+ source /setup_export_logs.sh
++ CLICKHOUSE_CI_LOGS_CREDENTIALS=/tmp/export-logs-config.sh
++ CLICKHOUSE_CI_LOGS_USER=ci
++ CLICKHOUSE_CI_LOGS_CLUSTER=system_logs_export
++ EXTRA_COLUMNS='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, '
++ EXTRA_COLUMNS_EXPRESSION='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id'
++ EXTRA_ORDER_BY_COLUMNS=check_name
++ EXTRA_COLUMNS_TRACE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), lines Array(LowCardinality(String)), '
++ EXTRA_COLUMNS_EXPRESSION_TRACE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayMap(x -> demangle(addressToSymbol(x)), trace)::Array(LowCardinality(String)) AS symbols, arrayMap(x -> addressToLine(x), trace)::Array(LowCardinality(String)) AS lines'
++ EXTRA_COLUMNS_COVERAGE_LOG='pull_request_number UInt32, commit_sha String, check_start_time DateTime('\''UTC'\''), check_name LowCardinality(String), instance_type LowCardinality(String), instance_id String, INDEX ix_pr (pull_request_number) TYPE set(100), INDEX ix_commit (commit_sha) TYPE set(100), INDEX ix_check_time (check_start_time) TYPE minmax, symbols Array(LowCardinality(String)), '
++ EXTRA_COLUMNS_EXPRESSION_COVERAGE_LOG='CAST(0 AS UInt32) AS pull_request_number, '\'''\'' AS commit_sha, now() AS check_start_time, toLowCardinality('\'''\'') AS check_name, toLowCardinality('\'''\'') AS instance_type, '\'''\'' AS instance_id, arrayDistinct(arrayMap(x -> demangle(addressToSymbol(x)), coverage))::Array(LowCardinality(String)) AS symbols'
+ source /repo/tests/docker_scripts/stress_tests.lib
++ OK='\tOK\t\N\t'
++ FAIL='\tFAIL\t\N\t'
++ FAILURE_CONTEXT_LINES=100
++ FAILURE_CONTEXT_MAX_LINE_WIDTH=300
+ dmesg --clear
+ set -e -x -a
+ USE_DATABASE_REPLICATED=0
+ USE_SHARED_CATALOG=0
++ rg -v '#' /usr/share/zoneinfo/zone.tab
++ awk '{print $3}'
++ shuf
++ head -n1
+ TZ=Asia/Anadyr
Chosen random timezone Asia/Anadyr
+ echo 'Chosen random timezone Asia/Anadyr'
+ ln -snf /usr/share/zoneinfo/Asia/Anadyr /etc/localtime
+ echo Asia/Anadyr
+ dpkg -i package_folder/clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-common-static.
(Reading database ... 48425 files and directories currently installed.)
Preparing to unpack .../clickhouse-common-static_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-common-static (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-common-static-dbg.
(Reading database ... 48452 files and directories currently installed.)
Preparing to unpack .../clickhouse-common-static-dbg_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-common-static-dbg (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-odbc-bridge.
(Reading database ... 48459 files and directories currently installed.)
Preparing to unpack .../clickhouse-odbc-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-odbc-bridge (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-library-bridge.
(Reading database ... 48465 files and directories currently installed.)
Preparing to unpack .../clickhouse-library-bridge_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-library-bridge (24.12.2.20221.altinityantalya+tsan) ...
+ dpkg -i package_folder/clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-server.
(Reading database ... 48471 files and directories currently installed.)
Preparing to unpack .../clickhouse-server_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-server (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-server (24.12.2.20221.altinityantalya+tsan) ...
ClickHouse binary is already located at /usr/bin/clickhouse
Symlink /usr/bin/clickhouse-server already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-server to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-client to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-local to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-benchmark to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-obfuscator to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-git-import to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-compressor to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-format to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-extract-from-config already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-extract-from-config to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-keeper already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-keeper to /usr/bin/clickhouse.
Symlink /usr/bin/clickhouse-keeper-converter already exists but it points to /clickhouse. Will replace the old symlink to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-keeper-converter to /usr/bin/clickhouse.
Creating symlink /usr/bin/clickhouse-disks to /usr/bin/clickhouse.
Creating symlink /usr/bin/ch to /usr/bin/clickhouse.
Creating symlink /usr/bin/chl to /usr/bin/clickhouse.
Creating symlink /usr/bin/chc to /usr/bin/clickhouse.
Creating clickhouse group if it does not exist.
groupadd -r clickhouse
Creating clickhouse user if it does not exist.
useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse clickhouse
Will set ulimits for clickhouse user in /etc/security/limits.d/clickhouse.conf.
Creating config directory /etc/clickhouse-server/config.d that is used for tweaks of main server configuration.
Creating config directory /etc/clickhouse-server/users.d that is used for tweaks of users configuration.
Config file /etc/clickhouse-server/config.xml already exists, will keep it and extract path info from it.
/etc/clickhouse-server/config.xml has /var/lib/clickhouse/ as data path.
/etc/clickhouse-server/config.xml has /var/log/clickhouse-server/ as log path.
Users config file /etc/clickhouse-server/users.xml already exists, will keep it and extract users info from it.
Log directory /var/log/clickhouse-server/ already exists.
Creating data directory /var/lib/clickhouse/.
Creating pid directory /var/run/clickhouse-server.
chown -R clickhouse:clickhouse '/var/log/clickhouse-server/'
chown -R clickhouse:clickhouse '/var/run/clickhouse-server'
chown clickhouse:clickhouse '/var/lib/clickhouse/'
groupadd -r clickhouse-bridge
useradd -r --shell /bin/false --home-dir /nonexistent -g clickhouse-bridge clickhouse-bridge
chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-odbc-bridge'
chown -R clickhouse-bridge:clickhouse-bridge '/usr/bin/clickhouse-library-bridge'
Password for the default user is an empty string. See /etc/clickhouse-server/users.xml and /etc/clickhouse-server/users.d to change it.
Setting capabilities for clickhouse binary. This is optional.
chown -R clickhouse:clickhouse '/etc/clickhouse-server'
ClickHouse has been successfully installed.
Start clickhouse-server with:
sudo clickhouse start
Start clickhouse-client with:
clickhouse-client
+ dpkg -i package_folder/clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb
Selecting previously unselected package clickhouse-client.
(Reading database ... 48488 files and directories currently installed.)
Preparing to unpack .../clickhouse-client_24.12.2.20221.altinityantalya+tsan_amd64.deb ...
Unpacking clickhouse-client (24.12.2.20221.altinityantalya+tsan) ...
Setting up clickhouse-client (24.12.2.20221.altinityantalya+tsan) ...
+ echo ''
+ [[ -z '' ]]
+ ch --query 'SELECT 1'
1
+ chl --query 'SELECT 1'
1
+ chc --version
ClickHouse client version 24.12.2.20221.altinityantalya (altinity build).
+ ln -sf /repo/tests/clickhouse-test /usr/bin/clickhouse-test
+ export CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py
+ CLICKHOUSE_GRPC_CLIENT=/repo/utils/grpc-client/clickhouse-grpc-client.py
+ source /repo/tests/docker_scripts/attach_gdb.lib
++ source /repo/tests/docker_scripts/utils.lib
+++ sysctl kernel.core_pattern=core.%e.%p-%P
kernel.core_pattern = core.%e.%p-%P
+++ sysctl fs.suid_dumpable=1
fs.suid_dumpable = 1
+ source /repo/tests/docker_scripts/utils.lib
++ sysctl kernel.core_pattern=core.%e.%p-%P
kernel.core_pattern = core.%e.%p-%P
++ sysctl fs.suid_dumpable=1
fs.suid_dumpable = 1
+ /repo/tests/config/install.sh
+ DEST_SERVER_PATH=/etc/clickhouse-server
+ DEST_CLIENT_PATH=/etc/clickhouse-client
+++ dirname /repo/tests/config/install.sh
++ cd /repo/tests/config
++ pwd -P
+ SRC_PATH=/repo/tests/config
+ '[' 0 -ge 2 ']'
+ FAST_TEST=0
+ NO_AZURE=0
+ [[ 0 -gt 0 ]]
+ echo 'Going to install test configs from /repo/tests/config into /etc/clickhouse-server'
+ mkdir -p /etc/clickhouse-server/config.d/
Going to install test configs from /repo/tests/config into /etc/clickhouse-server
+ mkdir -p /etc/clickhouse-server/users.d/
+ mkdir -p /etc/clickhouse-client
+ ln -sf /repo/tests/config/config.d/zookeeper_write.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/max_num_to_warn.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/listen.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/text_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/blob_storage_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/custom_settings_prefixes.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/database_catalog_drop_table_concurrency.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_access_control_improvements.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/macros.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/secure_ports.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/clusters.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/graphite.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/graphite_alternative.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/grpc_protocol.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/database_atomic.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/max_concurrent_queries.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree_settings.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/backoff_failed_mutation.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree_old_dirs_cleanup.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/test_cluster_with_incorrect_pw.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/keeper_port.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/logging_no_rotate.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/merge_tree.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/lost_forever_check.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/tcp_with_proxy.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/prometheus.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/top_level_domains_lists.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/top_level_domains_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/transactions.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/encryption.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/CORS.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/zookeeper_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/logger_trace.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/named_collection.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/ssl_certs.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/filesystem_cache_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/session_log.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/system_unfreeze.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_zero_copy_replication.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/nlp.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/forbidden_headers.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_keeper_map.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/custom_disks_base_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/display_name.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/compressed_marks_and_index.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/disable_s3_env_credentials.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/enable_wait_for_shutdown_replicated_tables.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/backups.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/filesystem_caches_path.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/validate_tcp_client_information.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/zero_copy_destructive_operations.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/block_number.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/handlers.yaml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/serverwide_trace_collector.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/rocksdb.xml /etc/clickhouse-server/config.d/
+ '[' 0 '!=' 1 ']'
+ ln -sf /repo/tests/config/config.d/legacy_geobase.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/users.d/log_queries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/readonly.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/access_management.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/database_atomic_drop_detach_sync.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/opentelemetry.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/remote_queries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/session_log_test.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/memory_profiler.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/no_fsync_metadata.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/filelog.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/enable_blobs_check.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/marks.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/insert_keeper_retries.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/prefetch_settings.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/nonconst_timezone.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/allow_introspection_functions.yaml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/replicated_ddl_entry.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/limits.yaml /etc/clickhouse-server/users.d/
+ [[ -n '' ]]
+ ln -sf /repo/tests/config/users.d/timeouts.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/ints_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/strings_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/decimals_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/executable_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/executable_pool_dictionary.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/test_function.xml /etc/clickhouse-server/
+ ln -sf /repo/tests/config/top_level_domains /etc/clickhouse-server/
+ ln -sf /repo/tests/config/regions_hierarchy.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/regions_names_en.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/regions_names_es.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/ext-en.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/ext-ru.txt /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/lem-en.bin /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/server.key /etc/clickhouse-server/
+ ln -sf /repo/tests/config/server.crt /etc/clickhouse-server/
+ ln -sf /repo/tests/config/dhparam.pem /etc/clickhouse-server/
+ ln -sf --backup=simple --suffix=_original.xml /repo/tests/config/config.d/query_masking_rules.xml /etc/clickhouse-server/config.d/
+ [[ -n '' ]]
+ rm -f /etc/clickhouse-server/config.d/zookeeper_fault_injection.xml
+ ln -sf /repo/tests/config/config.d/zookeeper.xml /etc/clickhouse-server/config.d/
+ [[ -n '' ]]
+ rm -f /etc/clickhouse-server/config.d/cannot_allocate_thread_injection.xml
+ value=1
+ sed --follow-symlinks -i 's|[01]|1|' /etc/clickhouse-server/config.d/keeper_port.xml
+ value=21907456
+ sed --follow-symlinks -i 's|[[:digit:]]\+|21907456|' /etc/clickhouse-server/config.d/keeper_port.xml
+ value=30001152
+ sed --follow-symlinks -i 's|[[:digit:]]\+|30001152|' /etc/clickhouse-server/config.d/keeper_port.xml
+ [[ -n '' ]]
+ [[ -n '' ]]
+ [[ '' == \1 ]]
+ [[ '' == \1 ]]
+ [[ 1 == \1 ]]
+ [[ 0 != \1 ]]
+ ln -sf /repo/tests/config/config.d/azure_storage_conf.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02944.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02963.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/config.d/storage_conf_02961.xml /etc/clickhouse-server/config.d/
+ ln -sf /repo/tests/config/users.d/s3_cache.xml /etc/clickhouse-server/users.d/
+ ln -sf /repo/tests/config/users.d/s3_cache_new.xml /etc/clickhouse-server/users.d/
+ [[ 0 == \1 ]]
+ ln -sf /repo/tests/config/client_config.xml /etc/clickhouse-client/config.xml
+ /repo/tests/docker_scripts/setup_minio.sh stateless
+ azurite-blob --blobHost 0.0.0.0 --blobPort 10000 --silent --inMemoryPersistence
+ export MINIO_ROOT_USER=clickhouse
+ MINIO_ROOT_USER=clickhouse
+ export MINIO_ROOT_PASSWORD=clickhouse
+ MINIO_ROOT_PASSWORD=clickhouse
+ main stateless
+ local query_dir
++ check_arg stateless
++ local query_dir
++ '[' '!' 1 -eq 1 ']'
++ case "$1" in
++ query_dir=0_stateless
++ echo 0_stateless
+ query_dir=0_stateless
+ '[' '!' -f ./minio ']'
+ start_minio
+ mkdir -p ./minio_data
+ ./minio --version
minio version RELEASE.2024-08-03T04-33-23Z (commit-id=6efb56851c40da88d1ca15112e2d686a4ecec6b3)
Runtime: go1.22.5 linux/amd64
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Copyright: 2015-2024 MinIO, Inc.
+ wait_for_it
+ local counter=0
+ local max_counter=60
+ ./minio server --address :11111 ./minio_data
+ local url=http://localhost:11111
+ params=('--silent' '--verbose')
+ local params
+ grep AccessDenied
+ curl --silent --verbose http://localhost:11111
+ [[ 0 == \6\0 ]]
+ echo 'trying to connect to minio'
+ sleep 1
trying to connect to minio
(node:273) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead.
(Use `node --trace-deprecation ...` to show where the warning was created)
In-memory extent storage is enabled with a limit of 15667.21 MB (16428259328 bytes, 50% of total memory).
Azurite Blob service is starting on 0.0.0.0:10000
INFO: Formatting 1st pool, 1 set(s), 1 drives per set.
INFO: WARNING: Host local has more than 0 drives of set. A host failure will result in data becoming unavailable.
Azurite Blob service successfully listens on http://0.0.0.0:10000
MinIO Object Storage Server
Copyright: 2015-2025 MinIO, Inc.
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64)
API: http://172.17.0.2:11111 http://127.0.0.1:11111
WebUI: http://172.17.0.2:36561 http://127.0.0.1:36561
Docs: https://min.io/docs/minio/linux/index.html
+ counter=1
+ curl --silent --verbose http://localhost:11111
+ grep AccessDenied
AccessDenied
Access Denied./18324FAAEDF18EA97dc7eb22d3288ec80374614e9088e31d3668a6922ead55932dd2a8e56373820f
+ lsof -i :11111
COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
minio 298 root 6u IPv6 27090 0t0 TCP *:11111 (LISTEN)
minio 298 root 9u IPv4 27089 0t0 TCP localhost:11111 (LISTEN)
minio 298 root 10u IPv6 27091 0t0 TCP localhost:11111 (LISTEN)
+ sleep 5
+ setup_minio stateless
+ local test_type=stateless
+ ./mc alias set clickminio http://localhost:11111 clickhouse clickhouse
Added `clickminio` successfully.
+ ./mc admin user add clickminio test testtest
Added user `test` successfully.
+ ./mc admin policy attach clickminio readwrite --user=test
Attached Policies: [readwrite]
To User: test
+ ./mc mb --ignore-existing clickminio/test
Bucket created successfully `clickminio/test`.
+ '[' stateless = stateless ']'
+ ./mc anonymous set public clickminio/test
Access permission for `clickminio/test` is set to `public`
+ upload_data 0_stateless /repo/tests/
+ local query_dir=0_stateless
+ local test_path=/repo/tests/
+ local data_path=/repo/tests//queries/0_stateless/data_minio
+ '[' -d /repo/tests//queries/0_stateless/data_minio ']'
+ ./mc cp --recursive /repo/tests//queries/0_stateless/data_minio/ clickminio/test/
`/repo/tests/queries/0_stateless/data_minio/03036_archive1.tar` -> `clickminio/test/03036_archive1.tar`
`/repo/tests/queries/0_stateless/data_minio/03036_archive2.zip` -> `clickminio/test/03036_archive2.zip`
`/repo/tests/queries/0_stateless/data_minio/02731.arrow` -> `clickminio/test/02731.arrow`
`/repo/tests/queries/0_stateless/data_minio/02876.parquet` -> `clickminio/test/02876.parquet`
`/repo/tests/queries/0_stateless/data_minio/02366_data.jsonl` -> `clickminio/test/02366_data.jsonl`
`/repo/tests/queries/0_stateless/data_minio/03036_archive1.zip` -> `clickminio/test/03036_archive1.zip`
`/repo/tests/queries/0_stateless/data_minio/02731.parquet` -> `clickminio/test/02731.parquet`
`/repo/tests/queries/0_stateless/data_minio/03036_archive2.tar` -> `clickminio/test/03036_archive2.tar`
`/repo/tests/queries/0_stateless/data_minio/03036_archive3.tar.gz` -> `clickminio/test/03036_archive3.tar.gz`
`/repo/tests/queries/0_stateless/data_minio/03036_compressed_file_archive.zip` -> `clickminio/test/03036_compressed_file_archive.zip`
`/repo/tests/queries/0_stateless/data_minio/03036_json_archive.zip` -> `clickminio/test/03036_json_archive.zip`
`/repo/tests/queries/0_stateless/data_minio/a.tsv` -> `clickminio/test/a.tsv`
`/repo/tests/queries/0_stateless/data_minio/b.tsv` -> `clickminio/test/b.tsv`
`/repo/tests/queries/0_stateless/data_minio/c.tsv` -> `clickminio/test/c.tsv`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Gordon/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/column1=Schmidt/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/column0=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/column0=Elizabeth/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/hive_partitioning/non_existing_column=Elizabeth/sample.parquet` -> `clickminio/test/hive_partitioning/non_existing_column=Elizabeth/sample.parquet`
`/repo/tests/queries/0_stateless/data_minio/json_data` -> `clickminio/test/json_data`
`/repo/tests/queries/0_stateless/data_minio/tsv_with_header.tsv` -> `clickminio/test/tsv_with_header.tsv`
Total: 5.42 MiB, Transferred: 5.42 MiB, Speed: 128.78 MiB/s
+ setup_aws_credentials
+ local minio_root_user=clickhouse
+ local minio_root_password=clickhouse
+ mkdir -p /root/.aws
+ cat
+ config_logs_export_cluster /etc/clickhouse-server/config.d/system_logs_export.yaml
+ set +x
File /tmp/export-logs-config.sh does not exist, do not setup
+ [[ -n '' ]]
+ export IS_FLAKY_CHECK=0
+ IS_FLAKY_CHECK=0
+ export NUM_TRIES
+ '[' 1 -gt 1 ']'
+ sudo -E -u clickhouse /usr/bin/clickhouse-server --config /etc/clickhouse-server/config.xml --daemon --pid-file /var/run/clickhouse-server/clickhouse-server.pid
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ sleep 1
+ for _ in {1..100}
+ clickhouse-client --query 'SELECT 1'
1
+ break
+ setup_logs_replication
+ set +x
File /tmp/export-logs-config.sh does not exist, do not setup
+ attach_gdb_to_clickhouse
++ run_with_retry 5 clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)'
++ [[ ahxB =~ e ]]
++ set_e=false
++ set +e
++ local total_retries=5
++ shift
++ local retry=0
++ '[' 0 -ge 5 ']'
++ clickhouse-client --query 'SELECT count() FROM system.build_options WHERE name = '\''CXX_FLAGS'\'' AND position('\''sanitize=address'\'' IN value)'
++ false
++ return
+ IS_ASAN=0
+ [[ 0 = \1 ]]
++ kill -l SIGRTMIN
+ RTMIN=34
+ echo '
set follow-fork-mode parent
handle SIGHUP nostop noprint pass
handle SIGINT nostop noprint pass
handle SIGQUIT nostop noprint pass
handle SIGPIPE nostop noprint pass
handle SIGTERM nostop noprint pass
handle SIGUSR1 nostop noprint pass
handle SIGUSR2 nostop noprint pass
handle SIG34 nostop noprint pass
info signals
continue
backtrace full
info registers
p top' 1 KiB of the 'stack:
p/x *(uint64_t[128]*)$sp
maintenance info sections
thread apply all backtrace full
disassemble /s
up
disassemble /s
up
disassemble /s
p "done"
detach
quit
'
+ sleep 5
+ ts '%Y-%m-%d %H:%M:%S'
++ cat /var/run/clickhouse-server/clickhouse-server.pid
+ gdb -batch -command script.gdb -p 435
aarch64-binfmt-P: Could not open '/lib/ld-linux-aarch64.so.1': No such file or directory
+ run_with_retry 60 clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\'''
+ [[ aehxB =~ e ]]
+ set_e=true
+ set +e
+ local total_retries=60
+ shift
+ local retry=0
+ '[' 0 -ge 60 ']'
+ clickhouse-client --query 'SELECT '\''Connected to clickhouse-server after attaching gdb'\'''
Connected to clickhouse-server after attaching gdb
+ true
+ set -e
+ return
+ clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_audit_logs
(
log JSON(time DateTime64(9))
)
ENGINE = MergeTree
ORDER BY tuple()'
+ clickhouse-client --allow_experimental_json_type=1 --query 'CREATE TABLE minio_server_logs
(
log JSON(time DateTime64(9))
)
ENGINE = MergeTree
ORDER BY tuple()'
+ ./mc admin config set clickminio logger_webhook:ch_server_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_server_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500
Successfully applied new settings.
+ ./mc admin config set clickminio audit_webhook:ch_audit_webhook 'endpoint=http://localhost:8123/?async_insert=1&wait_for_async_insert=0&async_insert_busy_timeout_min_ms=5000&async_insert_busy_timeout_max_ms=5000&async_insert_max_query_number=1000&async_insert_max_data_size=10485760&date_time_input_format=best_effort&query=INSERT%20INTO%20minio_audit_logs%20FORMAT%20JSONAsObject' queue_size=1000000 batch_size=500
Successfully applied new settings.
+ max_retries=100
+ retry=1
+ '[' 1 -le 100 ']'
+ echo 'clickminio restart attempt 1:'
clickminio restart attempt 1:
++ ./mc admin service restart clickminio --wait --json
++ jq -r .status
INFO: Restarting on service signal
MinIO Object Storage Server
Copyright: 2015-2025 MinIO, Inc.
License: GNU AGPLv3 - https://www.gnu.org/licenses/agpl-3.0.html
Version: RELEASE.2024-08-03T04-33-23Z (go1.22.5 linux/amd64)
API: http://172.17.0.2:11111 http://127.0.0.1:11111
WebUI: http://172.17.0.2:40083 http://127.0.0.1:40083
Docs: https://min.io/docs/minio/linux/index.html
Output of restart status: success
success
Restarted clickminio successfully.
+ output='success
success'
+ echo 'Output of restart status: success
success'
+ expected_output='success
success'
+ '[' 'success
success' = 'success
success' ']'
+ echo 'Restarted clickminio successfully.'
+ break
+ '[' 1 -gt 100 ']'
+ MC_ADMIN_PID=1348
+ ./mc admin trace clickminio
+ export -f run_tests
+ '[' 1 -gt 1 ']'
+ run_tests
+ set -x
+ read -ra ADDITIONAL_OPTIONS
+ HIGH_LEVEL_COVERAGE=YES
+ '[' 1 -gt 1 ']'
+ [[ -n '' ]]
+ [[ -n '' ]]
+ [[ 0 -eq 1 ]]
+ [[ '' -eq 1 ]]
+ [[ 0 -eq 1 ]]
++ clickhouse-client --query 'SELECT value LIKE '\''%SANITIZE_COVERAGE%'\'' FROM system.build_options WHERE name = '\''CXX_FLAGS'\'''
+ [[ 1 == 0 ]]
+ ADDITIONAL_OPTIONS+=('--jobs')
+ ADDITIONAL_OPTIONS+=('8')
+ [[ -n 1 ]]
+ [[ -n 4 ]]
+ ADDITIONAL_OPTIONS+=('--run-by-hash-num')
+ ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_NUM")
+ ADDITIONAL_OPTIONS+=('--run-by-hash-total')
+ ADDITIONAL_OPTIONS+=("$RUN_BY_HASH_TOTAL")
+ HIGH_LEVEL_COVERAGE=NO
+ [[ -n '' ]]
+ [[ NO = \Y\E\S ]]
+ ADDITIONAL_OPTIONS+=('--report-logs-stats')
+ try_run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ local total_retries=10
+ shift
+ fn_exists run_with_retry
+ declare -F run_with_retry
+ run_with_retry 10 clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ [[ aehxB =~ e ]]
+ set_e=true
+ set +e
+ local total_retries=10
+ shift
+ local retry=0
+ '[' 0 -ge 10 ']'
+ clickhouse-client -q 'insert into system.zookeeper (name, path, value) values ('\''auxiliary_zookeeper2'\'', '\''/test/chroot/'\'', '\'''\'')'
+ true
+ set -e
+ return
+ set +e
+ TEST_ARGS=(--testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries "/repo/tests/queries" --test-runs "$NUM_TRIES" "${ADDITIONAL_OPTIONS[@]}")
+ clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 1 --run-by-hash-total 4 --report-logs-stats
+ ts '%Y-%m-%d %H:%M:%S'
+ tee -a test_output/test_result.txt
2025-04-02 09:40:19 Using queries from '/repo/tests/queries' directory
2025-04-02 09:40:19 Connecting to ClickHouse server... OK
2025-04-02 09:40:19 Connected to server 24.12.2.20221.altinityantalya @ 82252d159dc02cab0f366aaa5691adc1545dd11d HEAD
2025-04-02 09:40:19 Found 1718 parallel tests and 139 sequential tests
2025-04-02 09:40:20 Running about 214 stateless tests (Process-3).
2025-04-02 09:40:20 01420_logical_functions_materialized_null: [ OK ] 0.74 sec.
2025-04-02 09:40:20 Running about 214 stateless tests (Process-5).
2025-04-02 09:40:20 03104_create_view_join: [ OK ] 0.74 sec.
2025-04-02 09:40:20 Running about 214 stateless tests (Process-6).
2025-04-02 09:40:20 03091_analyzer_same_table_name_in_different_databases: [ OK ] 0.89 sec.
2025-04-02 09:40:20 Running about 214 stateless tests (Process-9).
2025-04-02 09:40:20 02575_merge_prewhere_ephemeral: [ OK ] 1.19 sec.
2025-04-02 09:40:20 Running about 214 stateless tests (Process-7).
2025-04-02 09:40:20 02707_skip_index_with_in: [ OK ] 1.24 sec.
2025-04-02 09:40:21 Running about 214 stateless tests (Process-8).
2025-04-02 09:40:21 00873_t64_codec_date: [ OK ] 1.45 sec.
2025-04-02 09:40:21 01515_mv_and_array_join_optimisation_bag: [ OK ] 1.05 sec.
2025-04-02 09:40:21 Running about 214 stateless tests (Process-4).
2025-04-02 09:40:21 01213_alter_rename_column: [ OK ] 2.05 sec.
2025-04-02 09:40:21 02513_analyzer_duplicate_alias_crash_fix: [ OK ] 0.74 sec.
2025-04-02 09:40:22 03035_morton_encode_no_rows: [ OK ] 0.64 sec.
2025-04-02 09:40:22 01506_buffer_table_alter_block_structure_2: [ OK ] 0.89 sec.
2025-04-02 09:40:22 Running about 214 stateless tests (Process-10).
2025-04-02 09:40:22 00163_shard_join_with_empty_table: [ OK ] 3.15 sec.
2025-04-02 09:40:23 00576_nested_and_prewhere: [ OK ] 2.70 sec.
2025-04-02 09:40:23 01825_new_type_json_partitions: [ OK ] 0.74 sec.
2025-04-02 09:40:23 03149_analyzer_window_redefinition: [ OK ] 1.00 sec.
2025-04-02 09:40:23 00555_hasSubstr: [ OK ] 1.60 sec.
2025-04-02 09:40:23 01455_shard_leaf_max_rows_bytes_to_read: [ OK ] 2.66 sec.
2025-04-02 09:40:23 02363_mapupdate_improve: [ OK ] 1.09 sec.
2025-04-02 09:40:24 02662_sparse_columns_mutations_5: [ OK ] 0.89 sec.
2025-04-02 09:40:24 01867_support_datetime64_version_column: [ OK ] 1.09 sec.
2025-04-02 09:40:24 01533_optimize_skip_merged_partitions: [ OK ] 1.13 sec.
2025-04-02 09:40:24 02685_bson2: [ OK ] 0.74 sec.
2025-04-02 09:40:25 01761_round_year_bounds: [ OK ] 0.68 sec.
2025-04-02 09:40:26 02116_tuple_element_analyzer: [ OK ] 2.39 sec.
2025-04-02 09:40:27 00995_order_by_with_fill: [ OK ] 1.75 sec.
2025-04-02 09:40:28 02293_h3_distance: [ OK ] 1.25 sec.
2025-04-02 09:40:28 00534_functions_bad_arguments6: [ SKIPPED ] 0.00 sec.
2025-04-02 09:40:28 Reason: not running for current build
2025-04-02 09:40:29 00184_shard_distributed_group_by_no_merge: [ OK ] 3.09 sec.
2025-04-02 09:40:29 02541_lightweight_delete_on_cluster: [ OK ] 1.13 sec.
2025-04-02 09:40:30 03033_virtual_column_override: [ OK ] 0.84 sec.
2025-04-02 09:40:31 00089_group_by_arrays_of_fixed: [ OK ] 0.74 sec.
2025-04-02 09:40:32 01052_window_view_proc_tumble_to_now: [ OK ] 8.02 sec.
2025-04-02 09:40:32 00512_fractional_time_zones: [ OK ] 3.39 sec.
2025-04-02 09:40:33 02733_fix_distinct_in_order_bug_49622: [ OK ] 0.94 sec.
2025-04-02 09:40:34 02346_additional_filters_index: [ OK ] 2.08 sec.
2025-04-02 09:40:35 00301_csv: [ OK ] 14.34 sec.
2025-04-02 09:40:35 00983_summing_merge_tree_not_an_identifier: [ OK ] 0.74 sec.
2025-04-02 09:40:36 02239_bzip2_bug: [ OK ] 12.64 sec.
2025-04-02 09:40:36 02467_cross_join_three_table_functions: [ OK ] 0.69 sec.
2025-04-02 09:40:36 01852_jit_if: [ OK ] 1.03 sec.
2025-04-02 09:40:37 01537_fuzz_count_equal: [ OK ] 0.70 sec.
2025-04-02 09:40:37 00647_histogram_negative: [ OK ] 1.04 sec.
2025-04-02 09:40:37 02459_materialized_view_default_value: [ OK ] 1.16 sec.
2025-04-02 09:40:37 02374_regexp_replace: [ OK ] 0.70 sec.
2025-04-02 09:40:38 00206_empty_array_to_single: [ OK ] 0.93 sec.
2025-04-02 09:40:39 01070_h3_indexes_are_neighbors: [ OK ] 0.81 sec.
2025-04-02 09:40:39 00126_buffer: [ OK ] 2.65 sec.
2025-04-02 09:40:40 01903_csvwithnames_subset_of_columns: [ OK ] 16.14 sec.
2025-04-02 09:40:40 03208_groupArrayIntersect_serialization: [ OK ] 2.20 sec.
2025-04-02 09:40:41 01655_window_functions_null: [ OK ] 0.75 sec.
2025-04-02 09:40:41 03198_json_extract_more_types: [ OK ] 1.62 sec.
2025-04-02 09:40:42 03272_client_highlighting_bug: [ OK ] 2.87 sec.
2025-04-02 09:40:42 02041_conversion_between_date32_and_datetime64: [ OK ] 0.81 sec.
2025-04-02 09:40:44 02946_format_values: [ OK ] 3.21 sec.
2025-04-02 09:40:45 02558_system_processes_elapsed: [ OK ] 5.46 sec.
2025-04-02 09:40:45 03118_analyzer_multi_join_prewhere: [ OK ] 1.41 sec.
2025-04-02 09:40:46 01710_aggregate_projection_with_normalized_states: [ OK ] 3.54 sec.
2025-04-02 09:40:51 02933_paste_join: [ OK ] 4.81 sec.
2025-04-02 09:41:05 02985_shard_query_start_time: [ OK ] 22.07 sec.
2025-04-02 09:41:06 02481_low_cardinality_with_short_circuit_functins_mutations: [ OK ] 1.62 sec.
2025-04-02 09:41:07 02248_nullable_custom_types_to_string: [ OK ] 0.81 sec.
2025-04-02 09:41:09 01552_impl_aggfunc_cloneresize: [ OK ] 1.31 sec.
2025-04-02 09:41:10 00927_asof_joins: [ OK ] 1.77 sec.
2025-04-02 09:41:29 02534_keyed_siphash: [ OK ] 18.27 sec.
2025-04-02 09:41:29 02364_multiSearch_function_family: [ OK ] 43.55 sec.
2025-04-02 09:41:30 00516_is_inf_nan: [ OK ] 0.93 sec.
2025-04-02 09:41:31 03013_addDays_with_timezone: [ OK ] 1.98 sec.
2025-04-02 09:41:33 00736_disjunction_optimisation: [ OK ] 2.66 sec.
2025-04-02 09:41:35 03174_projection_deduplicate: [ OK ] 2.61 sec.
2025-04-02 09:41:36 02913_sum_map_state: [ OK ] 0.76 sec.
2025-04-02 09:41:39 02806_system_parts_columns_modification_time: [ OK ] 7.95 sec.
2025-04-02 09:41:40 02400_create_table_on_cluster_normalization: [ OK ] 3.92 sec.
2025-04-02 09:41:41 03198_dictionary_validate_primary_key_type: [ OK ] 1.24 sec.
2025-04-02 09:41:42 02725_null_group_key_with_rollup: [ OK ] 1.00 sec.
2025-04-02 09:41:44 01952_optimize_distributed_group_by_sharding_key: [ OK ] 2.17 sec.
2025-04-02 09:41:45 03125_analyzer_CTE_two_joins: [ OK ] 0.92 sec.
2025-04-02 09:41:48 02525_range_hashed_dictionary_update_field: [ OK ] 7.58 sec.
2025-04-02 09:41:48 01653_tuple_hamming_distance_2: [ OK ] 2.96 sec.
2025-04-02 09:41:49 02812_bug_with_unused_join_columns: [ OK ] 0.90 sec.
2025-04-02 09:41:52 01297_alter_distributed: [ OK ] 3.65 sec.
2025-04-02 09:41:53 03006_analyzer_executable_table_function: [ OK ] 0.91 sec.
2025-04-02 09:41:54 02833_array_join_columns: [ OK ] 1.56 sec.
2025-04-02 09:41:57 00398_url_functions: [ OK ] 7.94 sec.
2025-04-02 09:41:57 02434_cancel_insert_when_client_dies: [ SKIPPED ] 0.00 sec.
2025-04-02 09:41:57 Reason: not running for current build
2025-04-02 09:41:57 00626_in_syntax: [ OK ] 2.69 sec.
2025-04-02 09:41:58 02841_valid_json_and_xml_on_http_exception: [ OK ] 97.74 sec.
2025-04-02 09:41:58 02023_nullable_int_uint_where: [ OK ] 1.24 sec.
2025-04-02 09:41:59 02907_filter_pushdown_crash: [ OK ] 1.46 sec.
2025-04-02 09:42:00 02416_json_object_inference: [ OK ] 2.43 sec.
2025-04-02 09:42:02 01758_optimize_skip_unused_shards_once: [ OK ] 3.58 sec.
2025-04-02 09:42:02 02750_settings_alias_tcp_protocol: [ OK ] 2.12 sec.
2025-04-02 09:42:03 01117_greatest_least_case: [ OK ] 0.78 sec.
2025-04-02 09:42:03 00549_join_use_nulls: [ OK ] 1.28 sec.
2025-04-02 09:42:06 01656_join_defaul_enum: [ OK ] 2.66 sec.
2025-04-02 09:42:08 00678_murmurhash: [ OK ] 1.61 sec.
2025-04-02 09:42:08 01074_window_view_event_tumble_asc_join_populate: [ OK ] 9.06 sec.
2025-04-02 09:42:10 00917_least_sqr: [ OK ] 1.01 sec.
2025-04-02 09:42:15 01279_dist_group_by: [ OK ] 4.98 sec.
2025-04-02 09:42:16 00490_with_select: [ OK ] 0.98 sec.
2025-04-02 09:42:17 03279_pr_3_way_joins_inner_first: [ OK ] 9.64 sec.
2025-04-02 09:42:18 02366_with_fill_date: [ OK ] 0.85 sec.
2025-04-02 09:42:19 02996_nullable_arrayReduce: [ OK ] 2.80 sec.
2025-04-02 09:42:20 02498_analyzer_aggregate_functions_arithmetic_operations_pass_fix: [ OK ] 1.51 sec.
2025-04-02 09:42:22 00132_sets: [ OK ] 1.37 sec.
2025-04-02 09:42:37 01592_long_window_functions1: [ OK ] 112.24 sec.
2025-04-02 09:42:40 02916_date_text_parsing: [ OK ] 17.82 sec.
2025-04-02 09:42:42 02896_dwarf_format: [ OK ] 23.29 sec.
2025-04-02 09:42:43 00294_shard_enums: [ OK ] 5.07 sec.
2025-04-02 09:42:45 02242_join_rocksdb: [ OK ] 5.19 sec.
2025-04-02 09:42:45 02483_password_reset: [ OK ] 3.26 sec.
2025-04-02 09:42:46 00663_tiny_log_empty_insert: [ OK ] 0.85 sec.
2025-04-02 09:42:46 01943_log_column_sizes: [ OK ] 1.12 sec.
2025-04-02 09:42:48 02271_fix_column_matcher_and_column_transformer: [ OK ] 2.21 sec.
2025-04-02 09:42:50 02808_aliases_inside_case: [ OK ] 1.06 sec.
2025-04-02 09:42:50 00965_send_logs_level_concurrent_queries: [ OK ] 4.26 sec.
2025-04-02 09:42:52 03199_json_extract_dynamic: [ OK ] 1.83 sec.
2025-04-02 09:42:52 02677_analyzer_bitmap_has_any: [ OK ] 1.31 sec.
2025-04-02 09:42:52 02212_cte_and_table_alias: [ OK ] 0.79 sec.
2025-04-02 09:42:53 01669_join_or_duplicates: [ OK ] 1.25 sec.
2025-04-02 09:42:54 02676_distinct_reading_in_order_analyzer: [ OK ] 0.97 sec.
2025-04-02 09:42:54 00995_optimize_read_in_order_with_aggregation: [ OK ] 1.00 sec.
2025-04-02 09:42:54 01916_low_cardinality_interval: [ OK ] 0.85 sec.
2025-04-02 09:42:55 01455_time_zones: [ OK ] 0.82 sec.
2025-04-02 09:42:56 02025_dictionary_array_nested_map: [ OK ] 1.13 sec.
2025-04-02 09:42:56 00531_aggregate_over_nullable: [ OK ] 1.31 sec.
2025-04-02 09:42:58 01833_test_collation_alvarotuso: [ OK ] 1.16 sec.
2025-04-02 09:42:58 02481_aggregation_in_order_plan: [ OK ] 2.02 sec.
2025-04-02 09:42:59 00475_in_join_db_table: [ OK ] 1.65 sec.
2025-04-02 09:42:59 01072_nullable_jit: [ OK ] 1.18 sec.
2025-04-02 09:43:00 02698_marked_dropped_tables: [ OK ] 0.97 sec.
2025-04-02 09:43:03 03168_inconsistent_ast_formatting: [ OK ] 3.46 sec.
2025-04-02 09:43:04 00604_shard_remote_and_columns_with_defaults: [ OK ] 3.21 sec.
2025-04-02 09:43:04 02122_parallel_formatting_CSV: [ OK ] 21.70 sec.
2025-04-02 09:43:04 01634_uuid_fuzz: [ OK ] 1.35 sec.
2025-04-02 09:43:09 00500_point_in_polygon_bug_3_linestring_rotation_precision: [ OK ] 4.11 sec.
2025-04-02 09:43:11 02006_test_positional_arguments: [ OK ] 6.09 sec.
2025-04-02 09:43:11 00906_low_cardinality_cache: [ OK ] 140.85 sec.
2025-04-02 09:43:12 02534_s3_heap_use_after_free: [ OK ] 1.07 sec.
2025-04-02 09:43:13 03015_analyzer_groupby_fuzz_60772: [ OK ] 1.20 sec.
2025-04-02 09:43:14 01080_check_for_error_incorrect_size_of_nested_column: [ OK ] 2.01 sec.
2025-04-02 09:43:15 00557_array_resize: [ OK ] 1.37 sec.
2025-04-02 09:43:16 00957_format_with_clashed_aliases: [ OK ] 2.41 sec.
2025-04-02 09:43:19 02366_kql_func_string: [ OK ] 14.97 sec.
2025-04-02 09:43:19 02417_repeat_input_commands: [ OK ] 3.01 sec.
2025-04-02 09:43:20 02961_higher_order_constant_expressions: [ OK ] 1.11 sec.
2025-04-02 09:43:22 02377_majority_insert_quorum_zookeeper_long: [ OK ] 13.57 sec.
2025-04-02 09:43:23 00479_date_and_datetime_to_number: [ OK ] 0.96 sec.
2025-04-02 09:43:26 01891_partition_hash: [ OK ] 2.51 sec.
2025-04-02 09:43:27 03277_async_insert: [ OK ] 6.68 sec.
2025-04-02 09:43:28 02952_binary: [ OK ] 2.03 sec.
2025-04-02 09:43:30 00078_string_concat: [ OK ] 10.56 sec.
2025-04-02 09:43:30 02896_union_distinct_http_format: [ OK ] 2.57 sec.
2025-04-02 09:43:31 00537_quarters: [ OK ] 0.87 sec.
2025-04-02 09:43:32 00961_visit_param_buffer_underflow: [ OK ] 0.77 sec.
2025-04-02 09:43:33 00342_escape_sequences: [ OK ] 0.75 sec.
2025-04-02 09:43:34 02763_mutate_compact_part_with_skip_indices_and_projections: [ OK ] 5.52 sec.
2025-04-02 09:43:36 02098_hashed_array_dictionary_simple_key: [ OK ] 191.38 sec.
2025-04-02 09:43:36 01263_type_conversion_nvartolomei: [ OK ] 3.83 sec.
2025-04-02 09:43:37 02516_projections_and_context: [ OK ] 2.65 sec.
2025-04-02 09:43:37 01883_with_grouping_sets: [ OK ] 7.38 sec.
2025-04-02 09:43:37 02476_query_parameters_insert: [ OK ] 1.12 sec.
2025-04-02 09:43:38 02474_unhex_in_fix_string: [ OK ] 1.41 sec.
2025-04-02 09:43:39 01915_for_each_crakjie: [ OK ] 1.77 sec.
2025-04-02 09:43:40 00816_long_concurrent_alter_column: [ OK ] 24.70 sec.
2025-04-02 09:43:40 02026_describe_include_subcolumns: [ OK ] 1.16 sec.
2025-04-02 09:43:41 02500_prevent_drop_nested_if_empty_part: [ OK ] 2.58 sec.
2025-04-02 09:43:43 02513_parquet_orc_arrow_nullable_schema_inference: [ OK ] 5.85 sec.
2025-04-02 09:43:43 00976_asof_join_on: [ OK ] 5.30 sec.
2025-04-02 09:43:44 01495_subqueries_in_with_statement: [ OK ] 2.67 sec.
2025-04-02 09:43:44 01665_merge_tree_min_for_concurrent_read: [ OK ] 0.89 sec.
2025-04-02 09:43:44 01504_view_type_conversion: [ OK ] 0.89 sec.
2025-04-02 09:43:45 02691_drop_column_with_projections_replicated: [ OK ] 1.54 sec.
2025-04-02 09:43:46 01883_grouping_sets_crash: [ OK ] 1.74 sec.
2025-04-02 09:43:46 01710_projection_aggregation_in_order: [ OK ] 5.43 sec.
2025-04-02 09:43:47 01266_default_prewhere_reqq: [ OK ] 1.14 sec.
2025-04-02 09:43:47 00751_default_databasename_for_view: [ OK ] 1.64 sec.
2025-04-02 09:43:48 00905_field_with_aggregate_function_state: [ OK ] 0.79 sec.
2025-04-02 09:43:48 00702_join_with_using: [ OK ] 1.40 sec.
2025-04-02 09:43:49 01018_optimize_read_in_order_with_in_subquery: [ OK ] 0.73 sec.
2025-04-02 09:43:49 02137_mv_into_join: [ OK ] 1.14 sec.
2025-04-02 09:43:50 02018_multiple_with_fill_for_the_same_column: [ OK ] 0.64 sec.
2025-04-02 09:43:50 02790_optimize_skip_unused_shards_join: [ OK ] 1.29 sec.
2025-04-02 09:43:51 01320_optimize_skip_unused_shards_no_non_deterministic: [ OK ] 0.84 sec.
2025-04-02 09:43:51 02270_client_name: [ OK ] 7.28 sec.
2025-04-02 09:43:52 00414_time_zones_direct_conversion: [ OK ] 1.04 sec.
2025-04-02 09:43:52 03170_part_offset_as_table_column: [ OK ] 1.08 sec.
2025-04-02 09:43:53 01355_defaultValueOfArgumentType_bug: [ OK ] 0.64 sec.
2025-04-02 09:43:53 03020_order_by_SimpleAggregateFunction: [ OK ] 2.39 sec.
2025-04-02 09:43:53 03231_dynamic_uniq_group_by: [ OK ] 1.09 sec.
2025-04-02 09:43:54 02124_empty_uuid: [ OK ] 0.64 sec.
2025-04-02 09:43:54 03045_unknown_identifier_alias_substitution: [ OK ] 0.74 sec.
2025-04-02 09:43:54 01210_drop_view: [ OK ] 0.64 sec.
2025-04-02 09:43:54 02366_kql_extend: [ OK ] 1.65 sec.
2025-04-02 09:43:55 02313_dump_column_structure_low_cardinality: [ OK ] 0.64 sec.
2025-04-02 09:43:55 02366_union_decimal_conversion: [ OK ] 0.84 sec.
2025-04-02 09:43:56 00270_views_query_processing_stage: [ OK ] 0.84 sec.
2025-04-02 09:43:56 03204_index_hint_fuzzer: [ OK ] 0.64 sec.
2025-04-02 09:44:08 01099_parallel_distributed_insert_select: [ FAIL ] 207.46 sec.
2025-04-02 09:44:08 Reason: server died 32
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.084346 [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Logical error: 'Replica info is not initialized'.
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.160042 [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Stack trace (when copying this message, always include the lines below):
2025-04-02 09:44:08
2025-04-02 09:44:08 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x0000000020b03283
2025-04-02 09:44:08 1. ./build_docker/./src/Common/Exception.cpp:105: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dda74
2025-04-02 09:44:08 2. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade
2025-04-02 09:44:08 3. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba
2025-04-02 09:44:08 4. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c
2025-04-02 09:44:08 5. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e
2025-04-02 09:44:08 6. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d
2025-04-02 09:44:08 7. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe
2025-04-02 09:44:08 8. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b
2025-04-02 09:44:08 9. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d
2025-04-02 09:44:08 10. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47
2025-04-02 09:44:08 11. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c
2025-04-02 09:44:08 12. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc
2025-04-02 09:44:08 13. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa
2025-04-02 09:44:08 14. ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: void std::__function::__policy_invoker::__call_impl::ThreadFromGlobalPoolImpl(DB::CompletedPipelineExecutor::execute()::$_0&&)::'lambda'(), void ()>>(std::__function::__policy_storage const*) @ 0x000000001c52036a
2025-04-02 09:44:08 15. ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ? @ 0x00000000104ef7ee
2025-04-02 09:44:08 16. ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ? @ 0x00000000104f79dc
2025-04-02 09:44:08 17. __tsan_thread_start_func @ 0x0000000007b08428
2025-04-02 09:44:08 18. ? @ 0x00007fd7c201fac3
2025-04-02 09:44:08 19. ? @ 0x00007fd7c20b1850
2025-04-02 09:44:08
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.164119 [ 2683 ] BaseDaemon: ########################################
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.165045 [ 2683 ] BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2170) (query_id: 5495cad3-cad2-435c-8e6e-a56e3e787891) (query: INSERT INTO distributed_01099_b SELECT * FROM urlCluster('test_cluster_two_shards', 'http://localhost:8123/?query=select+{1,2,3}+format+TSV', 'TSV', 's String');) Received signal Aborted (6)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.165988 [ 2683 ] BaseDaemon:
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.166563 [ 2683 ] BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x000055701ef34a3c 0x000055701ef31d6e 0x000055701ef33c0d 0x00005570235a9bfe 0x0000557023089f8b 0x00005570235a984d 0x00005570230aea47 0x000055702309e54c 0x000055702309d8cc 0x000055702309d5aa 0x000055702309c36a 0x000055701706b7ee 0x00005570170739dc 0x000055700e684428 0x00007fd7c201fac3 0x00007fd7c20b1850
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.369611 [ 2683 ] BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.369941 [ 2683 ] BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:35.566929 [ 2683 ] BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:41.719899 [ 2683 ] BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:58.543986 [ 2683 ] BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:58.550350 [ 2683 ] BaseDaemon: 4. ? @ 0x00007fd7c1fcd520
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:58.552001 [ 2683 ] BaseDaemon: 5. ? @ 0x00007fd7c20219fd
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:58.552734 [ 2683 ] BaseDaemon: 6. ? @ 0x00007fd7c1fcd476
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:40:58.552991 [ 2683 ] BaseDaemon: 7. ? @ 0x00007fd7c1fb37f3
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:18.915371 [ 2683 ] BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:19.730227 [ 2683 ] BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:19.946326 [ 2683 ] BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:19.949528 [ 2683 ] BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:40.626177 [ 2683 ] BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:41:57.988457 [ 2683 ] BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:02.121691 [ 2683 ] BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:02.526694 [ 2683 ] BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:02.984820 [ 2683 ] BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:03.568477 [ 2683 ] BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:03.880034 [ 2683 ] BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:03.988427 [ 2683 ] BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:04.260448 [ 2683 ] BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:04.269401 [ 2683 ] BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:05.376232 [ 2683 ] BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.230286 [ 2683 ] BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.231304 [ 2683 ] BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.373556 [ 2683 ] BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.671125 [ 2683 ] BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.675053 [ 2683 ] BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.675906 [ 2683 ] BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.676393 [ 2683 ] BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.678745 [ 2683 ] BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.680546 [ 2683 ] BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.681280 [ 2683 ] BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.681932 [ 2683 ] BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.682611 [ 2683 ] BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:06.685281 [ 2683 ] BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.228552 [ 2683 ] BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.231516 [ 2683 ] BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.231787 [ 2683 ] BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.591536 [ 2683 ] BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.595342 [ 2683 ] BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*, 2ul>(std::tuple>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>&, std::__tuple_indices<2ul>)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:07.596206 [ 2683 ] BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x00000000104f79dc
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:24.747786 [ 2683 ] BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:24.749564 [ 2683 ] BaseDaemon: 27. ? @ 0x00007fd7c201fac3
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:24.749778 [ 2683 ] BaseDaemon: 28. ? @ 0x00007fd7c20b1850
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:42:37.541566 [ 2683 ] BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:43:54.468595 [ 2683 ] BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.
2025-04-02 09:44:08 [faf538363b2d] 2025.04.02 09:43:54.470745 [ 2683 ] BaseDaemon: Changed settings: min_compress_block_size = 1147731, max_compress_block_size = 2885936, max_block_size = 62186, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 98860, max_insert_threads = 2, max_threads = 1, max_read_buffer_size = 995864, connect_timeout_with_failover_ms = 2000, connect_timeout_with_failover_secure_ms = 3000, idle_connection_timeout = 36000, s3_max_get_rps = 1000000, s3_max_get_burst = 2000000, s3_max_put_rps = 1000000, s3_max_put_burst = 2000000, s3_check_objects_after_upload = true, use_uncompressed_cache = true, max_remote_read_network_bandwidth = 1000000000000, max_remote_write_network_bandwidth = 1000000000000, max_local_read_bandwidth = 1000000000000, max_local_write_bandwidth = 1000000000000, stream_like_engine_allow_direct_select = true, enable_multiple_prewhere_read_steps = false, replication_wait_for_inactive_replica_timeout = 30, min_count_to_compile_expression = 0, compile_aggregate_expressions = false, min_count_to_compile_aggregate_expression = 0, group_by_two_level_threshold = 881876, group_by_two_level_threshold_bytes = 13928924, enable_memory_bound_merging_of_aggregation_results = false, allow_nonconst_timezone_arguments = true, parallel_distributed_insert_select = 1, input_format_parallel_parsing = false, min_chunk_bytes_for_parallel_parsing = 17686868, merge_tree_coarse_index_granularity = 26, min_bytes_to_use_direct_io = 10737418240, min_bytes_to_use_mmap_io = 1, log_queries = true, insert_quorum_timeout = 60000, merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability = 0.5199999809265137, http_wait_end_of_query = true, http_response_buffer_size = 3622091, fsync_metadata = false, http_send_timeout = 60., http_receive_timeout = 60., use_index_for_in_with_subqueries_max_values = 1000000000, opentelemetry_start_trace_probability = 0.10000000149011612, max_rows_to_read = 20000000, max_bytes_to_read = 1000000000000, max_bytes_to_read_leaf = 1000000000000, max_rows_to_group_by = 10000000000, max_bytes_before_external_group_by = 10737418240, max_rows_to_sort = 10000000000, max_bytes_to_sort = 10000000000, prefer_external_sort_block_bytes = 0, max_bytes_before_external_sort = 10737418240, max_bytes_before_remerge_sort = 1309929691, max_result_rows = 1000000000, max_result_bytes = 1000000000, max_execution_time = 600., max_execution_time_leaf = 600., max_execution_speed = 100000000000, max_execution_speed_bytes = 10000000000000, timeout_before_checking_execution_speed = 300., max_estimated_execution_time = 600., max_columns_to_read = 20000, max_temporary_columns = 20000, max_temporary_non_const_columns = 20000, max_rows_in_set = 10000000000, max_bytes_in_set = 10000000000, max_rows_in_join = 10000000000, max_bytes_in_join = 10000000000, cross_join_min_rows_to_compress = 1, cross_join_min_bytes_to_compress = 0, max_rows_to_transfer = 1000000000, max_bytes_to_transfer = 1000000000, max_rows_in_distinct = 10000000000, max_bytes_in_distinct = 10000000000, max_memory_usage = 5000000000, max_memory_usage_for_user = 32000000000, max_untracked_memory = 1048576, memory_profiler_step = 1048576, max_network_bandwidth = 100000000000, max_network_bytes = 1000000000000, max_network_bandwidth_for_user = 100000000000, max_network_bandwidth_for_all_users = 100000000000, max_temporary_data_on_disk_size_for_user = 100000000000, max_temporary_data_on_disk_size_for_query = 100000000000, max_backup_bandwidth = 100000000000, log_comment = '01099_parallel_distributed_insert_select.sql', send_logs_level = 'error', prefer_localhost_replica = false, aggregation_in_order_max_block_bytes = 42302895, read_in_order_two_level_merge_threshold = 83, max_hyperscan_regexp_length = 1000000, max_hyperscan_regexp_total_length = 10000000, allow_introspection_functions = true, database_atomic_wait_for_drop_and_detach_synchronously = true, optimize_if_chain_to_multiif = true, optimize_if_transform_strings_to_enum = true, optimize_append_index = true, query_cache_max_size_in_bytes = 10000000, query_cache_max_entries = 100000, distributed_ddl_entry_format_version = 6, external_storage_max_read_rows = 10000000000, external_storage_max_read_bytes = 10000000000, remote_filesystem_read_method = 'read', merge_tree_min_bytes_per_task_for_remote_reading = 16777216, merge_tree_compact_parts_min_granules_to_multibuffer_read = 101, async_insert_busy_timeout_max_ms = 5000, enable_filesystem_cache = true, enable_filesystem_cache_on_write_operations = true, throw_on_error_from_cache_on_write_operations = true, filesystem_cache_segments_batch_size = 3, use_page_cache_for_disks_without_file_cache = true, page_cache_inject_eviction = true, load_marks_asynchronously = true, allow_prefetched_read_pool_for_remote_filesystem = true, allow_prefetched_read_pool_for_local_filesystem = false, filesystem_prefetch_step_marks = 50, filesystem_prefetch_max_memory_usage = 67108864, filesystem_prefetches_limit = 10, max_streams_for_merge_tree_reading = 1000, optimize_sorting_by_input_stream_properties = false, insert_keeper_max_retries = 100, insert_keeper_retry_initial_backoff_ms = 1, insert_keeper_retry_max_backoff_ms = 10, insert_keeper_fault_injection_probability = 0.009999999776482582, optimize_distinct_in_order = false, allow_experimental_parallel_reading_from_replicas = 0, session_timezone = 'Asia/Anadyr', output_format_native_write_json_as_string = true
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: INSERT INTO distributed_01099_b SELECT * FROM urlCluster('test_cluster_two_shards', 'http://localhost:8123/?query=select+{1,2,3}+format+TSV', 'TSV', 's String');)
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08 parallel_distributed_insert_select=1
2025-04-02 09:44:08 test_shard_localhost
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 test_cluster_two_shards_localhost
2025-04-02 09:44:08 0 2
2025-04-02 09:44:08 1 2
2025-04-02 09:44:08 2 2
2025-04-02 09:44:08 test_cluster_two_shards
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 local
2025-04-02 09:44:08 0 2
2025-04-02 09:44:08 1 2
2025-04-02 09:44:08 2 2
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 0 4
2025-04-02 09:44:08 1 4
2025-04-02 09:44:08 2 4
2025-04-02 09:44:08 test_cluster_1_shard_3_replicas_1_unavailable
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 local
2025-04-02 09:44:08 test_cluster_1_shard_3_replicas_1_unavailable with storageCluster
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08 parallel_distributed_insert_select=1
2025-04-02 09:44:08 test_shard_localhost
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 test_cluster_two_shards_localhost
2025-04-02 09:44:08 0 2
2025-04-02 09:44:08 1 2
2025-04-02 09:44:08 2 2
2025-04-02 09:44:08 test_cluster_two_shards
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 local
2025-04-02 09:44:08 0 2
2025-04-02 09:44:08 1 2
2025-04-02 09:44:08 2 2
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 0 4
2025-04-02 09:44:08 1 4
2025-04-02 09:44:08 2 4
2025-04-02 09:44:08 test_cluster_1_shard_3_replicas_1_unavailable
2025-04-02 09:44:08 distributed
2025-04-02 09:44:08 local
2025-04-02 09:44:08 test_cluster_1_shard_3_replicas_1_unavailable with storageCluster
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 881876 --group_by_two_level_threshold_bytes 13928924 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 1 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 17686868 --max_read_buffer_size 995864 --prefer_localhost_replica 0 --max_block_size 62186 --max_joined_block_size_rows 98860 --max_threads 1 --optimize_append_index 1 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 0 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 83 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 42302895 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method pread_threadpool --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 3 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 26 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 1309929691 --min_compress_block_size 1147731 --max_compress_block_size 2885936 --merge_tree_compact_parts_min_granules_to_multibuffer_read 101 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 3622091 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Asia/Anadyr --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.52 --prefer_external_sort_block_bytes 0 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 0 --max_parsing_threads 0 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 10737418240 --vertical_merge_algorithm_min_rows_to_activate 202301 --vertical_merge_algorithm_min_columns_to_activate 51 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 9457502 --merge_max_block_size 13313 --index_granularity 2497 --min_bytes_for_wide_part 1007520646 --marks_compress_block_size 44173 --primary_key_compress_block_size 76970 --replace_long_file_name_to_hash 1 --max_file_name_length 18 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 112788977 --compact_parts_max_granules_to_buffer 1 --compact_parts_merge_max_bytes_to_prefetch_part 26611372 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 45 --old_parts_lifetime 466 --prewarm_mark_cache 0 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 0 --prewarm_primary_key_cache 1
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_r8zyw55s
2025-04-02 09:44:08 01037_polygon_dicts_correctness_fast: [ FAIL ] 205.22 sec.
2025-04-02 09:44:08 Reason: return code: 1
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: select 'dictGet', 'dict_array' as dict_name, tuple(x, y) as key,
2025-04-02 09:44:08 dictGet(dict_name, 'value', key) from points order by x, y;)
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08 Files /repo/tests/queries/0_stateless/01037_polygon_dicts_correctness_fast.ans and /repo/tests/queries/0_stateless/default/tmp/resultsPOLYGON_INDEX_CELL.out differ
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08 Files /repo/tests/queries/0_stateless/01037_polygon_dicts_correctness_fast.ans and /repo/tests/queries/0_stateless/default/tmp/resultsPOLYGON_INDEX_CELL.out differ
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 1000000 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 1 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 5180581 --max_read_buffer_size 1037881 --prefer_localhost_replica 0 --max_block_size 87894 --max_joined_block_size_rows 56140 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 100 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 48680710 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method io_uring --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 0 --merge_tree_coarse_index_granularity 9 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 377248284 --min_compress_block_size 2503184 --max_compress_block_size 1677278 --merge_tree_compact_parts_min_granules_to_multibuffer_read 63 --optimize_sorting_by_input_stream_properties 1 --http_response_buffer_size 7310187 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 0 --session_timezone Africa/Khartoum --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.62 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 0 --max_parsing_threads 1 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 0 --query_plan_join_swap_table false --output_format_native_write_json_as_string 0 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.8343553042906489 --prefer_fetch_merged_part_size_threshold 1 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 1 --index_granularity_bytes 9389802 --merge_max_block_size 7933 --index_granularity 32877 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 44211 --primary_key_compress_block_size 28850 --replace_long_file_name_to_hash 0 --max_file_name_length 0 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 108251988 --compact_parts_max_granules_to_buffer 168 --compact_parts_merge_max_bytes_to_prefetch_part 19307120 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 44 --old_parts_lifetime 431 --prewarm_mark_cache 0 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 0 --use_primary_key_cache 0 --prewarm_primary_key_cache 0
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_jj3cad05
2025-04-02 09:44:08 Process Process-8:
2025-04-02 09:44:08 Traceback (most recent call last):
2025-04-02 09:44:08 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:08 self.run()
2025-04-02 09:44:08 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:08 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:08 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:08 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:08 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:08 stop_tests()
2025-04-02 09:44:08 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:08 cleanup_child_processes(os.getpid())
2025-04-02 09:44:08 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:08 child_pgid = os.getpgid(child)
2025-04-02 09:44:08 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:08 Child processes of 1396:
2025-04-02 09:44:08 8321 /bin/sh -c pgrep --parent 1396 -a
2025-04-02 09:44:08 01174_select_insert_isolation: [ FAIL ] 4.20 sec.
2025-04-02 09:44:08 Reason: return code: 1
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: CREATE TEMPORARY TABLE tmp AS SELECT arraySort(groupArray(n)), arraySort(groupArray(m)), arraySort(groupArray(_part)) FROM mt FORMAT Null;)
2025-04-02 09:44:08 Error on processing query: Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:38054): while receiving packet from localhost:9000. (NETWORK_ERROR) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: INSERT INTO mt VALUES /* (1, 2) */ (1, 2);)
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: INSERT INTO mt VALUES /* (42, 3) */ (42, 3);)
2025-04-02 09:44:08 Error on processing query: Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: BEGIN TRANSACTION;)
2025-04-02 09:44:08 /repo/tests/queries/0_stateless/01174_select_insert_isolation.sh: line 57: kill: (8239) - No such process
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 2 --group_by_two_level_threshold 583253 --group_by_two_level_threshold_bytes 1 --distributed_aggregation_memory_efficient 1 --fsync_metadata 1 --output_format_parallel_formatting 1 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 10994245 --max_read_buffer_size 758883 --prefer_localhost_replica 1 --max_block_size 18438 --max_joined_block_size_rows 85998 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 0 --optimize_or_like_chain 1 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 63 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 18076765 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 1 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method read --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 5 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 1Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 4 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 1731213327 --min_compress_block_size 2282938 --max_compress_block_size 2800166 --merge_tree_compact_parts_min_granules_to_multibuffer_read 114 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 10436828 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Africa/Juba --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.6 --prefer_external_sort_block_bytes 100000000 --cross_join_min_rows_to_compress 1 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 0 --max_parsing_threads 10 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.06328676173441272 --prefer_fetch_merged_part_size_threshold 10737418240 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 8936507819 --index_granularity_bytes 8016494 --merge_max_block_size 7681 --index_granularity 22884 --min_bytes_for_wide_part 1073741824 --marks_compress_block_size 75442 --primary_key_compress_block_size 85352 --replace_long_file_name_to_hash 0 --max_file_name_length 62 --min_bytes_for_full_part_storage 536870912 --compact_parts_max_bytes_to_buffer 256693951 --compact_parts_max_granules_to_buffer 256 --compact_parts_merge_max_bytes_to_prefetch_part 15649468 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_aitya9fp
2025-04-02 09:44:08 00290_shard_aggregation_memory_efficient: [ FAIL ] 3.06 sec.
2025-04-02 09:44:08 Reason: return code: 32
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: SELECT concat(toString(number), arrayStringConcat(arrayMap(x -> '.', range(number % 10)))) AS k FROM remote('127.0.0.{2,3}', currentDatabase(), numbers_10_00290) WHERE number < (randConstant() % 2 ? 4999 : 10000) GROUP BY k ORDER BY k LIMIT 10;)
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08 0
2025-04-02 09:44:08 1.
2025-04-02 09:44:08 10
2025-04-02 09:44:08 100
2025-04-02 09:44:08 1000
2025-04-02 09:44:08 1001.
2025-04-02 09:44:08 1002..
2025-04-02 09:44:08 1003...
2025-04-02 09:44:08 1004....
2025-04-02 09:44:08 1005.....
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 1 --group_by_two_level_threshold 871323 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 0 --fsync_metadata 1 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 8378908 --max_read_buffer_size 940545 --prefer_localhost_replica 1 --max_block_size 23467 --max_joined_block_size_rows 23883 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 0 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 27 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 43441359 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 1 --min_bytes_to_use_mmap_io 835472481 --local_filesystem_read_method pread_threadpool --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 10 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 17 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 91107026 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 1636820905 --min_compress_block_size 2920666 --max_compress_block_size 2766304 --merge_tree_compact_parts_min_granules_to_multibuffer_read 37 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 4948613 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 0 --session_timezone Asia/Anadyr --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.06 --prefer_external_sort_block_bytes 100000000 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 0 --max_parsing_threads 10 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 1 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.4775148015149031 --prefer_fetch_merged_part_size_threshold 6105471224 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 100 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 42482410 --index_granularity_bytes 16223784 --merge_max_block_size 4141 --index_granularity 35355 --min_bytes_for_wide_part 969947172 --marks_compress_block_size 73824 --primary_key_compress_block_size 31923 --replace_long_file_name_to_hash 1 --max_file_name_length 11 --min_bytes_for_full_part_storage 346230518 --compact_parts_max_bytes_to_buffer 499646739 --compact_parts_max_granules_to_buffer 175 --compact_parts_merge_max_bytes_to_prefetch_part 12372579 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 0 --old_parts_lifetime 10 --prewarm_mark_cache 1 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 1 --use_primary_key_cache 1 --prewarm_primary_key_cache 1
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_p8mmsbpt
2025-04-02 09:44:08 03094_grouparraysorted_memory: [ FAIL ] 13.25 sec.
2025-04-02 09:44:08 Reason: return code: 32
2025-04-02 09:44:08 Error on processing query: Code: 32. DB::Exception: Attempt to read after eof: while receiving packet from localhost:9000. (ATTEMPT_TO_READ_AFTER_EOF) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: INSERT INTO 03094_grouparrysorted_src SELECT * FROM generateRandom() LIMIT 500000;)
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 3 --group_by_two_level_threshold 62841 --group_by_two_level_threshold_bytes 3050572 --distributed_aggregation_memory_efficient 0 --fsync_metadata 1 --output_format_parallel_formatting 0 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 10671017 --max_read_buffer_size 892159 --prefer_localhost_replica 1 --max_block_size 28880 --max_joined_block_size_rows 53918 --max_threads 3 --optimize_append_index 0 --optimize_if_chain_to_multiif 1 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 1 --optimize_or_like_chain 1 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 0 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 13318541 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 7151100800 --local_filesystem_read_method io_uring --remote_filesystem_read_method read --local_filesystem_read_prefetch 1 --filesystem_cache_segments_batch_size 0 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 0 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 64Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 8Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 0 --merge_tree_coarse_index_granularity 21 --optimize_distinct_in_order 0 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 9510112074 --max_bytes_before_remerge_sort 2936926746 --min_compress_block_size 587895 --max_compress_block_size 2677723 --merge_tree_compact_parts_min_granules_to_multibuffer_read 17 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 776315 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 0 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Africa/Khartoum --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.59 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 1 --max_parsing_threads 0 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 0 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 10737418240 --vertical_merge_algorithm_min_rows_to_activate 1000000 --vertical_merge_algorithm_min_columns_to_activate 1 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 20293203 --merge_max_block_size 634 --index_granularity 46379 --min_bytes_for_wide_part 0 --marks_compress_block_size 97035 --primary_key_compress_block_size 87914 --replace_long_file_name_to_hash 0 --max_file_name_length 95 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 90193327 --compact_parts_max_granules_to_buffer 256 --compact_parts_merge_max_bytes_to_prefetch_part 21482907 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 1 --use_primary_key_cache 0 --prewarm_primary_key_cache 1
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_5akil08g
2025-04-02 09:44:08 03173_set_transformed_partition_pruning: [ FAIL ] 115.58 sec.
2025-04-02 09:44:08 Reason: return code: 210
2025-04-02 09:44:08 Error on processing query: Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:40734): while receiving packet from localhost:9000. (NETWORK_ERROR) (version 24.12.2.20221.altinityantalya (altinity build))
2025-04-02 09:44:08 (query: SYSTEM FLUSH LOGS;)
2025-04-02 09:44:08 , result:
2025-04-02 09:44:08
2025-04-02 09:44:08 -- Single partition by function
2025-04-02 09:44:08 0
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- Nested partition by function
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, LowCardinality
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, Nullable
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, LowCardinality + Nullable
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Non-safe cast
2025-04-02 09:44:08 2
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- Multiple partition columns
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- LowCardinality set
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nullable set
2025-04-02 09:44:08 1
2025-04-02 09:44:08
2025-04-02 09:44:08 stdout:
2025-04-02 09:44:08 -- Single partition by function
2025-04-02 09:44:08 0
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- Nested partition by function
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, LowCardinality
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, Nullable
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nested partition by function, LowCardinality + Nullable
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Non-safe cast
2025-04-02 09:44:08 2
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- Multiple partition columns
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 2
2025-04-02 09:44:08 -- LowCardinality set
2025-04-02 09:44:08 1
2025-04-02 09:44:08 1
2025-04-02 09:44:08 -- Nullable set
2025-04-02 09:44:08 1
2025-04-02 09:44:08
2025-04-02 09:44:08
2025-04-02 09:44:08 Settings used in the test: --max_insert_threads 1 --group_by_two_level_threshold 616796 --group_by_two_level_threshold_bytes 50000000 --distributed_aggregation_memory_efficient 1 --fsync_metadata 1 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 15003412 --max_read_buffer_size 1014831 --prefer_localhost_replica 0 --max_block_size 23721 --max_joined_block_size_rows 41929 --max_threads 3 --optimize_append_index 0 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 1 --optimize_read_in_order 1 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 29 --optimize_aggregation_in_order 0 --aggregation_in_order_max_block_bytes 29778262 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 10737418240 --min_bytes_to_use_mmap_io 7716725907 --local_filesystem_read_method io_uring --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 2 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 50 --filesystem_prefetch_step_bytes 100Mi --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 30 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 10737418240 --max_bytes_before_remerge_sort 2111304381 --min_compress_block_size 2493443 --max_compress_block_size 2031853 --merge_tree_compact_parts_min_granules_to_multibuffer_read 79 --optimize_sorting_by_input_stream_properties 1 --http_response_buffer_size 9527763 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 0 --session_timezone America/Hermosillo --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.47 --prefer_external_sort_block_bytes 0 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 100000000 --min_external_table_block_size_bytes 0 --max_parsing_threads 1 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table false --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-02 09:44:08
2025-04-02 09:44:08 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 0.0 --prefer_fetch_merged_part_size_threshold 1 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 15 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 10737418240 --index_granularity_bytes 15484493 --merge_max_block_size 18545 --index_granularity 47548 --min_bytes_for_wide_part 0 --marks_compress_block_size 96724 --primary_key_compress_block_size 25258 --replace_long_file_name_to_hash 0 --max_file_name_length 0 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 98090925 --compact_parts_max_granules_to_buffer 191 --compact_parts_merge_max_bytes_to_prefetch_part 25923386 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 0 --old_parts_lifetime 10 --prewarm_mark_cache 0 --use_const_adaptive_granularity 1 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 1
2025-04-02 09:44:08
2025-04-02 09:44:08 Database: test_r3v5e7gu
2025-04-02 09:44:09 00754_distributed_optimize_skip_select_on_unused_shards_with_prewhere: [ FAIL ] 18.98 sec.
2025-04-02 09:44:09 Reason: return code: 210
2025-04-02 09:44:09 Code: 210. DB::NetException: Connection reset by peer, while reading from socket (peer: [::1]:9000, local: [::1]:38092): (localhost:9000, ::1, local address: [::1]:38092). (NETWORK_ERROR)
2025-04-02 09:44:09
2025-04-02 09:44:09 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-02 09:44:09
2025-04-02 09:44:09 , result:
2025-04-02 09:44:09
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 1
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 0
2025-04-02 09:44:09 1
2025-04-02 09:44:09 4
2025-04-02 09:44:09 4
2025-04-02 09:44:09 2
2025-04-02 09:44:09 4
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09
2025-04-02 09:44:09 stdout:
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 1
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 0
2025-04-02 09:44:09 1
2025-04-02 09:44:09 4
2025-04-02 09:44:09 4
2025-04-02 09:44:09 2
2025-04-02 09:44:09 4
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09 OK
2025-04-02 09:44:09
2025-04-02 09:44:09
2025-04-02 09:44:09 Settings used in the test: --max_insert_threads 1 --group_by_two_level_threshold 1 --group_by_two_level_threshold_bytes 12376090 --distributed_aggregation_memory_efficient 1 --fsync_metadata 0 --output_format_parallel_formatting 0 --input_format_parallel_parsing 0 --min_chunk_bytes_for_parallel_parsing 14696045 --max_read_buffer_size 974646 --prefer_localhost_replica 0 --max_block_size 85150 --max_joined_block_size_rows 72309 --max_threads 2 --optimize_append_index 0 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 1 --optimize_or_like_chain 0 --optimize_substitute_columns 0 --enable_multiple_prewhere_read_steps 0 --read_in_order_two_level_merge_threshold 21 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 32961818 --use_uncompressed_cache 0 --min_bytes_to_use_direct_io 1663439070 --min_bytes_to_use_mmap_io 1 --local_filesystem_read_method io_uring --remote_filesystem_read_method threadpool --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 5 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 0 --remote_filesystem_read_prefetch 1 --allow_prefetched_read_pool_for_remote_filesystem 1 --filesystem_prefetch_max_memory_usage 128Mi --filesystem_prefetches_limit 0 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 1 --compile_sort_description 1 --merge_tree_coarse_index_granularity 17 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 10737418240 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 1663352174 --min_compress_block_size 1688596 --max_compress_block_size 1626321 --merge_tree_compact_parts_min_granules_to_multibuffer_read 51 --optimize_sorting_by_input_stream_properties 0 --http_response_buffer_size 7030236 --http_wait_end_of_query True --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 0 --min_count_to_compile_aggregate_expression 0 --min_count_to_compile_sort_description 3 --session_timezone Mexico/BajaSur --use_page_cache_for_disks_without_file_cache True --page_cache_inject_eviction True --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.91 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 0 --cross_join_min_bytes_to_compress 0 --min_external_table_block_size_bytes 0 --max_parsing_threads 10 --optimize_functions_to_subcolumns 0 --parallel_replicas_local_plan 1 --query_plan_join_swap_table true --output_format_native_write_json_as_string 1 --enable_vertical_final 0
2025-04-02 09:44:09
2025-04-02 09:44:09 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 1 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 75 --allow_vertical_merges_from_compact_to_wide_parts 1 --min_merge_bytes_to_use_direct_io 1 --index_granularity_bytes 2836420 --merge_max_block_size 2374 --index_granularity 6538 --min_bytes_for_wide_part 0 --marks_compress_block_size 68763 --primary_key_compress_block_size 21565 --replace_long_file_name_to_hash 1 --max_file_name_length 128 --min_bytes_for_full_part_storage 284675622 --compact_parts_max_bytes_to_buffer 31544265 --compact_parts_max_granules_to_buffer 1 --compact_parts_merge_max_bytes_to_prefetch_part 1445265 --cache_populated_by_fetch 1 --concurrent_part_removal_threshold 100 --old_parts_lifetime 156 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 1
2025-04-02 09:44:09
2025-04-02 09:44:09 Database: test_e9rkp3xh
2025-04-02 09:44:10 02916_joinget_dependency: [ FAIL ] 3.46 sec.
2025-04-02 09:44:10 Reason: return code: 210
2025-04-02 09:44:10 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-02 09:44:10
2025-04-02 09:44:10 Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
2025-04-02 09:44:10
2025-04-02 09:44:10 , result:
2025-04-02 09:44:10
2025-04-02 09:44:10 0
2025-04-02 09:44:10
2025-04-02 09:44:10 stdout:
2025-04-02 09:44:10 0
2025-04-02 09:44:10
2025-04-02 09:44:10
2025-04-02 09:44:10 Settings used in the test: --max_insert_threads 3 --group_by_two_level_threshold 497490 --group_by_two_level_threshold_bytes 1 --distributed_aggregation_memory_efficient 1 --fsync_metadata 1 --output_format_parallel_formatting 1 --input_format_parallel_parsing 1 --min_chunk_bytes_for_parallel_parsing 10545029 --max_read_buffer_size 690725 --prefer_localhost_replica 0 --max_block_size 96554 --max_joined_block_size_rows 80874 --max_threads 3 --optimize_append_index 1 --optimize_if_chain_to_multiif 0 --optimize_if_transform_strings_to_enum 0 --optimize_read_in_order 1 --optimize_or_like_chain 0 --optimize_substitute_columns 1 --enable_multiple_prewhere_read_steps 1 --read_in_order_two_level_merge_threshold 42 --optimize_aggregation_in_order 1 --aggregation_in_order_max_block_bytes 27500700 --use_uncompressed_cache 1 --min_bytes_to_use_direct_io 7120490215 --min_bytes_to_use_mmap_io 10737418240 --local_filesystem_read_method io_uring --remote_filesystem_read_method read --local_filesystem_read_prefetch 0 --filesystem_cache_segments_batch_size 100 --read_from_filesystem_cache_if_exists_otherwise_bypass_cache 1 --throw_on_error_from_cache_on_write_operations 1 --remote_filesystem_read_prefetch 0 --allow_prefetched_read_pool_for_remote_filesystem 0 --filesystem_prefetch_max_memory_usage 32Mi --filesystem_prefetches_limit 10 --filesystem_prefetch_min_bytes_for_single_read_task 16Mi --filesystem_prefetch_step_marks 0 --filesystem_prefetch_step_bytes 0 --compile_aggregate_expressions 0 --compile_sort_description 1 --merge_tree_coarse_index_granularity 32 --optimize_distinct_in_order 1 --max_bytes_before_external_sort 1289811384 --max_bytes_before_external_group_by 0 --max_bytes_before_remerge_sort 402439281 --min_compress_block_size 2682903 --max_compress_block_size 2551524 --merge_tree_compact_parts_min_granules_to_multibuffer_read 60 --optimize_sorting_by_input_stream_properties 1 --http_response_buffer_size 3180472 --http_wait_end_of_query False --enable_memory_bound_merging_of_aggregation_results 1 --min_count_to_compile_expression 3 --min_count_to_compile_aggregate_expression 3 --min_count_to_compile_sort_description 3 --session_timezone Asia/Anadyr --use_page_cache_for_disks_without_file_cache False --page_cache_inject_eviction False --merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability 0.19 --prefer_external_sort_block_bytes 1 --cross_join_min_rows_to_compress 100000000 --cross_join_min_bytes_to_compress 1 --min_external_table_block_size_bytes 0 --max_parsing_threads 1 --optimize_functions_to_subcolumns 1 --parallel_replicas_local_plan 0 --query_plan_join_swap_table auto --output_format_native_write_json_as_string 1 --enable_vertical_final 1
2025-04-02 09:44:10
2025-04-02 09:44:10 MergeTree settings used in test: --ratio_of_defaults_for_sparse_serialization 1.0 --prefer_fetch_merged_part_size_threshold 434423316 --vertical_merge_algorithm_min_rows_to_activate 1 --vertical_merge_algorithm_min_columns_to_activate 11 --allow_vertical_merges_from_compact_to_wide_parts 0 --min_merge_bytes_to_use_direct_io 1535311281 --index_granularity_bytes 1107511 --merge_max_block_size 8924 --index_granularity 15032 --min_bytes_for_wide_part 0 --marks_compress_block_size 45698 --primary_key_compress_block_size 40129 --replace_long_file_name_to_hash 1 --max_file_name_length 128 --min_bytes_for_full_part_storage 0 --compact_parts_max_bytes_to_buffer 22309324 --compact_parts_max_granules_to_buffer 1 --compact_parts_merge_max_bytes_to_prefetch_part 12585697 --cache_populated_by_fetch 0 --concurrent_part_removal_threshold 100 --old_parts_lifetime 480 --prewarm_mark_cache 1 --use_const_adaptive_granularity 0 --enable_index_granularity_compression 0 --use_primary_key_cache 1 --prewarm_primary_key_cache 0
2025-04-02 09:44:10
2025-04-02 09:44:10 Database: test_uok0f9lx
2025-04-02 09:44:10 02995_index_4: [ SKIPPED ] 0.00 sec.
2025-04-02 09:44:10 Reason: not running for current build
2025-04-02 09:44:18 03173_forbid_qualify: [ FAIL ] 0.00 sec.
2025-04-02 09:44:18 Reason: server died
2025-04-02 09:44:18 Server does not respond to health check
2025-04-02 09:44:18
2025-04-02 09:44:18 Process Process-6:
2025-04-02 09:44:18 Traceback (most recent call last):
2025-04-02 09:44:18 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:18 self.run()
2025-04-02 09:44:18 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:18 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:18 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:18 stop_tests()
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:18 cleanup_child_processes(os.getpid())
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:18 child_pgid = os.getpgid(child)
2025-04-02 09:44:18 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:18 Child processes of 1392:
2025-04-02 09:44:18 8335 /bin/sh -c pgrep --parent 1392 -a
2025-04-02 09:44:18 00801_daylight_saving_time_hour_underflow: [ FAIL ] 0.00 sec.
2025-04-02 09:44:18 Reason: server died
2025-04-02 09:44:18 Server does not respond to health check
2025-04-02 09:44:18
2025-04-02 09:44:18 03032_string_to_variant_cast: [ FAIL ] 0.00 sec.
2025-04-02 09:44:18 Reason: server died
2025-04-02 09:44:18 Server does not respond to health check
2025-04-02 09:44:18
2025-04-02 09:44:18 Process Process-3:
2025-04-02 09:44:18 Traceback (most recent call last):
2025-04-02 09:44:18 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:18 self.run()
2025-04-02 09:44:18 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:18 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:18 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:18 stop_tests()
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:18 cleanup_child_processes(os.getpid())
2025-04-02 09:44:18 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:18 child_pgid = os.getpgid(child)
2025-04-02 09:44:18 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:18 Child processes of 1388:
2025-04-02 09:44:18 8341 /bin/sh -c pgrep --parent 1388 -a
2025-04-02 09:44:18 02706_arrow_different_dictionaries: [ FAIL ] 0.00 sec.
2025-04-02 09:44:18 Reason: server died
2025-04-02 09:44:18 Server does not respond to health check
2025-04-02 09:44:18
2025-04-02 09:44:18 02891_array_shingles: [ FAIL ] 0.00 sec.
2025-04-02 09:44:18 Reason: server died
2025-04-02 09:44:18 Server does not respond to health check
2025-04-02 09:44:18
2025-04-02 09:44:19 Process Process-7:
2025-04-02 09:44:19 Traceback (most recent call last):
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:19 self.run()
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:19 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:19 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:19 stop_tests()
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:19 cleanup_child_processes(os.getpid())
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:19 child_pgid = os.getpgid(child)
2025-04-02 09:44:19 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:19 Child processes of 1395:
2025-04-02 09:44:19 8343 /bin/sh -c pgrep --parent 1395 -a
2025-04-02 09:44:19 Process Process-5:
2025-04-02 09:44:19 Process Process-10:
2025-04-02 09:44:19 Traceback (most recent call last):
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:19 self.run()
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:19 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:19 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:19 stop_tests()
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:19 cleanup_child_processes(os.getpid())
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:19 child_pgid = os.getpgid(child)
2025-04-02 09:44:19 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:19 Child processes of 1390:
2025-04-02 09:44:19 8352 /bin/sh -c pgrep --parent 1390 -a
2025-04-02 09:44:19 Traceback (most recent call last):
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:19 self.run()
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:19 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:19 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:19 stop_tests()
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:19 cleanup_child_processes(os.getpid())
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:19 child_pgid = os.getpgid(child)
2025-04-02 09:44:19 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:19 Child processes of 1404:
2025-04-02 09:44:19 8351 /bin/sh -c pgrep --parent 1404 -a
2025-04-02 09:44:19 00907_set_index_max_rows: [ FAIL ] 0.00 sec.
2025-04-02 09:44:19 Reason: server died
2025-04-02 09:44:19 Server does not respond to health check
2025-04-02 09:44:19
2025-04-02 09:44:19 Process Process-9:
2025-04-02 09:44:19 Traceback (most recent call last):
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:19 self.run()
2025-04-02 09:44:19 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:19 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:19 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:19 stop_tests()
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:19 cleanup_child_processes(os.getpid())
2025-04-02 09:44:19 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:19 child_pgid = os.getpgid(child)
2025-04-02 09:44:19 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:19 Child processes of 1401:
2025-04-02 09:44:19 8365 /bin/sh -c pgrep --parent 1401 -a
2025-04-02 09:44:20 01801_dateDiff_DateTime64: [ FAIL ] 0.00 sec.
2025-04-02 09:44:20 Reason: server died
2025-04-02 09:44:20 Server does not respond to health check
2025-04-02 09:44:20
2025-04-02 09:44:20 Process Process-4:
2025-04-02 09:44:20 Traceback (most recent call last):
2025-04-02 09:44:20 File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
2025-04-02 09:44:20 self.run()
2025-04-02 09:44:20 File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run
2025-04-02 09:44:20 self._target(*self._args, **self._kwargs)
2025-04-02 09:44:20 File "/usr/bin/clickhouse-test", line 2626, in run_tests_process
2025-04-02 09:44:20 return run_tests_array(*args, **kwargs)
2025-04-02 09:44:20 File "/usr/bin/clickhouse-test", line 2407, in run_tests_array
2025-04-02 09:44:20 stop_tests()
2025-04-02 09:44:20 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:44:20 cleanup_child_processes(os.getpid())
2025-04-02 09:44:20 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:44:20 child_pgid = os.getpgid(child)
2025-04-02 09:44:20 ProcessLookupError: [Errno 3] No such process
2025-04-02 09:44:20 Child processes of 1389:
2025-04-02 09:44:20 8371 /bin/sh -c pgrep --parent 1389 -a
2025-04-02 09:57:09 Hung check failed: [Errno 111] Connection refused
2025-04-02 09:57:09 Server died, terminating all processes...
2025-04-02 09:57:09 Running 139 stateless tests (MainProcess).
2025-04-02 09:57:09 Child processes of 1372:
2025-04-02 09:57:09 1375 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 1 --run-by-hash-total 4 --report-logs-stats
2025-04-02 09:57:09 1381 python3 /usr/bin/clickhouse-test --testname --shard --zookeeper --check-zookeeper-session --hung-check --print-time --no-drop-if-fail --capture-client-stacktrace --queries /repo/tests/queries --test-runs 1 --hung-check --print-time --jobs 8 --run-by-hash-num 1 --run-by-hash-total 4 --report-logs-stats
2025-04-02 09:57:09 8379 /bin/sh -c pgrep --parent 1372 -a
2025-04-02 09:57:09 Traceback (most recent call last):
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 3716, in
2025-04-02 09:57:09 main(args)
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 3063, in main
2025-04-02 09:57:09 total_tests_run += do_run_tests(args.jobs, test_suite)
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 2697, in do_run_tests
2025-04-02 09:57:09 run_tests_array(
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 2343, in run_tests_array
2025-04-02 09:57:09 stop_tests()
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 452, in stop_tests
2025-04-02 09:57:09 cleanup_child_processes(os.getpid())
2025-04-02 09:57:09 File "/usr/bin/clickhouse-test", line 440, in cleanup_child_processes
2025-04-02 09:57:09 child_pgid = os.getpgid(child)
2025-04-02 09:57:09 ProcessLookupError: [Errno 3] No such process
+ set -e
+ echo 'Files in current directory'
+ ls -la ./
Files in current directory
total 127164
drwxr-xr-x 1 root root 4096 Apr 2 09:40 .
drwxr-xr-x 1 root root 4096 Apr 2 09:40 ..
lrwxrwxrwx 1 root root 7 Sep 12 2024 bin -> usr/bin
drwxr-xr-x 2 root root 4096 Apr 18 2022 boot
drwxr-xr-x 14 root root 3840 Apr 2 09:39 dev
-rwxr-xr-x 1 root root 0 Apr 2 09:39 .dockerenv
drwxr-xr-x 1 root root 4096 Apr 2 09:39 etc
drwxr-xr-x 2 root root 4096 Apr 18 2022 home
lrwxrwxrwx 1 root root 7 Sep 12 2024 lib -> usr/lib
lrwxrwxrwx 1 root root 9 Sep 12 2024 lib32 -> usr/lib32
lrwxrwxrwx 1 root root 9 Sep 12 2024 lib64 -> usr/lib64
lrwxrwxrwx 1 root root 10 Sep 12 2024 libx32 -> usr/libx32
-rwxr-xr-x 1 root root 26927256 Jan 16 09:53 mc
drwxr-xr-x 2 root root 4096 Sep 12 2024 media
-rwxr-xr-x 1 root root 103174296 Jan 16 09:53 minio
drwxr-xr-x 4 root root 4096 Apr 2 09:40 minio_data
drwxr-xr-x 2 root root 4096 Sep 12 2024 mnt
drwxr-xr-x 1 root root 4096 Jan 15 09:08 opt
-rw-r--r-- 1 root root 0 Feb 15 2024 .package-cache-mutate
drwxrwxr-x 2 1000 1000 4096 Apr 2 09:39 package_folder
dr-xr-xr-x 307 root root 0 Apr 2 09:39 proc
drwxrwxr-x 17 1000 1000 4096 Apr 2 09:36 repo
-rw-rw-r-- 1 root root 863 Jan 16 09:27 requirements.txt
drwx------ 1 root root 4096 Apr 2 09:40 root
drwxr-xr-x 1 root root 4096 Apr 2 09:39 run
lrwxrwxrwx 1 root root 8 Sep 12 2024 sbin -> usr/sbin
-rw-r--r-- 1 root root 747 Apr 2 09:40 script.gdb
-rwxrwxr-x 1 root root 10851 Jan 15 09:07 setup_export_logs.sh
drwxr-xr-x 2 root root 4096 Sep 12 2024 srv
dr-xr-xr-x 13 root root 0 Apr 2 09:39 sys
drwxrwxr-x 2 1000 1000 4096 Apr 2 09:40 test_output
drwxrwxrwt 1 root root 4096 Apr 2 09:57 tmp
drwxr-xr-x 1 root root 4096 Sep 12 2024 usr
drwxr-xr-x 1 root root 4096 Sep 12 2024 var
Files in root directory
+ echo 'Files in root directory'
+ ls -la /
total 127164
drwxr-xr-x 1 root root 4096 Apr 2 09:40 .
drwxr-xr-x 1 root root 4096 Apr 2 09:40 ..
lrwxrwxrwx 1 root root 7 Sep 12 2024 bin -> usr/bin
drwxr-xr-x 2 root root 4096 Apr 18 2022 boot
drwxr-xr-x 14 root root 3840 Apr 2 09:39 dev
-rwxr-xr-x 1 root root 0 Apr 2 09:39 .dockerenv
drwxr-xr-x 1 root root 4096 Apr 2 09:39 etc
drwxr-xr-x 2 root root 4096 Apr 18 2022 home
lrwxrwxrwx 1 root root 7 Sep 12 2024 lib -> usr/lib
lrwxrwxrwx 1 root root 9 Sep 12 2024 lib32 -> usr/lib32
lrwxrwxrwx 1 root root 9 Sep 12 2024 lib64 -> usr/lib64
lrwxrwxrwx 1 root root 10 Sep 12 2024 libx32 -> usr/libx32
-rwxr-xr-x 1 root root 26927256 Jan 16 09:53 mc
drwxr-xr-x 2 root root 4096 Sep 12 2024 media
-rwxr-xr-x 1 root root 103174296 Jan 16 09:53 minio
drwxr-xr-x 4 root root 4096 Apr 2 09:40 minio_data
drwxr-xr-x 2 root root 4096 Sep 12 2024 mnt
drwxr-xr-x 1 root root 4096 Jan 15 09:08 opt
-rw-r--r-- 1 root root 0 Feb 15 2024 .package-cache-mutate
drwxrwxr-x 2 1000 1000 4096 Apr 2 09:39 package_folder
dr-xr-xr-x 307 root root 0 Apr 2 09:39 proc
drwxrwxr-x 17 1000 1000 4096 Apr 2 09:36 repo
-rw-rw-r-- 1 root root 863 Jan 16 09:27 requirements.txt
drwx------ 1 root root 4096 Apr 2 09:40 root
drwxr-xr-x 1 root root 4096 Apr 2 09:39 run
lrwxrwxrwx 1 root root 8 Sep 12 2024 sbin -> usr/sbin
-rw-r--r-- 1 root root 747 Apr 2 09:40 script.gdb
-rwxrwxr-x 1 root root 10851 Jan 15 09:07 setup_export_logs.sh
drwxr-xr-x 2 root root 4096 Sep 12 2024 srv
dr-xr-xr-x 13 root root 0 Apr 2 09:39 sys
drwxrwxr-x 2 1000 1000 4096 Apr 2 09:40 test_output
drwxrwxrwt 1 root root 4096 Apr 2 09:57 tmp
drwxr-xr-x 1 root root 4096 Sep 12 2024 usr
drwxr-xr-x 1 root root 4096 Sep 12 2024 var
+ clickhouse-client -q 'system flush logs'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ stop_logs_replication
+ echo 'Detach all logs replication'
Detach all logs replication
+ clickhouse-client --query 'select database||'\''.'\''||table from system.tables where database = '\''system'\'' and (table like '\''%_sender'\'' or table like '\''%_watcher'\'')'
+ tee /dev/stderr
+ timeout --preserve-status --signal TERM --kill-after 5m 15m xargs -n1 -r -i clickhouse-client --query 'drop table {}'
xargs: warning: options --max-args and --replace/-I/-i are mutually exclusive, ignoring previous --max-args value
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ logs_saver_client_options='--max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0'
+ failed_to_save_logs=0
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_log into outfile '\''/test_output/query_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.zookeeper_log into outfile '\''/test_output/zookeeper_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.trace_log into outfile '\''/test_output/trace_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.transactions_info_log into outfile '\''/test_output/transactions_info_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.metric_log into outfile '\''/test_output/metric_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.blob_storage_log into outfile '\''/test_output/blob_storage_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.error_log into outfile '\''/test_output/error_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'select * from system.query_metric_log into outfile '\''/test_output/query_metric_log.tsv.zst'\'' format TSVWithNamesAndTypes'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ failed_to_save_logs=1
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ sleep 1
+ clickhouse-client -q 'SYSTEM FLUSH ASYNC INSERT QUEUE'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_audit_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_audit_logs.jsonl.zst'\'' FORMAT JSONEachRow'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ clickhouse-client --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 -q 'SELECT log FROM minio_server_logs ORDER BY log.time INTO OUTFILE '\''/test_output/minio_server_logs.jsonl.zst'\'' FORMAT JSONEachRow'
Code: 210. DB::NetException: Connection refused (localhost:9000). (NETWORK_ERROR)
+ :
+ sudo clickhouse stop
==8436==Installed the sigaction for signal 11
==8436==Installed the sigaction for signal 7
==8436==Installed the sigaction for signal 8
==8436==Using llvm-symbolizer found at: /usr/bin/llvm-symbolizer-19
***** Running under ThreadSanitizer v3 (pid 8436) *****
ThreadSanitizer: growing sync allocator: 0 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 0 out of 262144*4096
ThreadSanitizer: growing heap block allocator: 1 out of 262144*4096
ThreadSanitizer: growing sync allocator: 1 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 2 out of 262144*4096
ThreadSanitizer: growing sync allocator: 2 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 3 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 4 out of 1048576*1024
ThreadSanitizer: growing heap block allocator: 3 out of 262144*4096
ThreadSanitizer: growing sync allocator: 5 out of 1048576*1024
ThreadSanitizer: growing sync allocator: 6 out of 1048576*1024
/var/run/clickhouse-server/clickhouse-server.pid file exists and contains pid = 435.
The process with pid = 435 does not exist.
Stats: SizeClassAllocator64: 10M mapped (1M rss) in 20235 allocations; remains 5797
01 ( 16): mapped: 256K allocs: 1024 frees: 768 inuse: 256 num_freed_chunks 16128 avail: 16384 rss: 16K releases: 0 last released: 0K region: 0x720400000000
02 ( 32): mapped: 256K allocs: 7168 frees: 6656 inuse: 512 num_freed_chunks 7680 avail: 8192 rss: 176K releases: 0 last released: 0K region: 0x720800000000
03 ( 48): mapped: 256K allocs: 5632 frees: 4608 inuse: 1024 num_freed_chunks 4437 avail: 5461 rss: 252K releases: 0 last released: 0K region: 0x720c00000000
04 ( 64): mapped: 256K allocs: 1792 frees: 1408 inuse: 384 num_freed_chunks 3712 avail: 4096 rss: 108K releases: 0 last released: 0K region: 0x721000000000
05 ( 80): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 3020 avail: 3276 rss: 28K releases: 0 last released: 0K region: 0x721400000000
06 ( 96): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 2474 avail: 2730 rss: 20K releases: 0 last released: 0K region: 0x721800000000
07 ( 112): mapped: 256K allocs: 384 frees: 128 inuse: 256 num_freed_chunks 2084 avail: 2340 rss: 36K releases: 0 last released: 0K region: 0x721c00000000
08 ( 128): mapped: 256K allocs: 640 frees: 384 inuse: 256 num_freed_chunks 1792 avail: 2048 rss: 80K releases: 0 last released: 0K region: 0x722000000000
09 ( 144): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1692 avail: 1820 rss: 8K releases: 0 last released: 0K region: 0x722400000000
10 ( 160): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1510 avail: 1638 rss: 8K releases: 0 last released: 0K region: 0x722800000000
11 ( 176): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1361 avail: 1489 rss: 4K releases: 0 last released: 0K region: 0x722c00000000
12 ( 192): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1237 avail: 1365 rss: 16K releases: 0 last released: 0K region: 0x723000000000
13 ( 208): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 1132 avail: 1260 rss: 4K releases: 0 last released: 0K region: 0x723400000000
14 ( 224): mapped: 256K allocs: 512 frees: 256 inuse: 256 num_freed_chunks 914 avail: 1170 rss: 92K releases: 0 last released: 0K region: 0x723800000000
15 ( 240): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 836 avail: 1092 rss: 40K releases: 0 last released: 0K region: 0x723c00000000
16 ( 256): mapped: 256K allocs: 256 frees: 0 inuse: 256 num_freed_chunks 768 avail: 1024 rss: 40K releases: 0 last released: 0K region: 0x724000000000
17 ( 320): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 691 avail: 819 rss: 16K releases: 0 last released: 0K region: 0x724400000000
18 ( 384): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 554 avail: 682 rss: 8K releases: 0 last released: 0K region: 0x724800000000
19 ( 448): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 457 avail: 585 rss: 8K releases: 0 last released: 0K region: 0x724c00000000
20 ( 512): mapped: 256K allocs: 128 frees: 0 inuse: 128 num_freed_chunks 384 avail: 512 rss: 4K releases: 0 last released: 0K region: 0x725000000000
21 ( 640): mapped: 256K allocs: 102 frees: 0 inuse: 102 num_freed_chunks 307 avail: 409 rss: 4K releases: 0 last released: 0K region: 0x725400000000
22 ( 768): mapped: 256K allocs: 85 frees: 0 inuse: 85 num_freed_chunks 256 avail: 341 rss: 4K releases: 0 last released: 0K region: 0x725800000000
23 ( 896): mapped: 256K allocs: 73 frees: 0 inuse: 73 num_freed_chunks 219 avail: 292 rss: 8K releases: 0 last released: 0K region: 0x725c00000000
24 ( 1024): mapped: 256K allocs: 64 frees: 0 inuse: 64 num_freed_chunks 192 avail: 256 rss: 8K releases: 0 last released: 0K region: 0x726000000000
25 ( 1280): mapped: 256K allocs: 204 frees: 102 inuse: 102 num_freed_chunks 102 avail: 204 rss: 252K releases: 0 last released: 0K region: 0x726400000000
26 ( 1536): mapped: 256K allocs: 42 frees: 0 inuse: 42 num_freed_chunks 128 avail: 170 rss: 4K releases: 0 last released: 0K region: 0x726800000000
27 ( 1792): mapped: 256K allocs: 36 frees: 0 inuse: 36 num_freed_chunks 110 avail: 146 rss: 4K releases: 0 last released: 0K region: 0x726c00000000
28 ( 2048): mapped: 256K allocs: 32 frees: 0 inuse: 32 num_freed_chunks 96 avail: 128 rss: 8K releases: 0 last released: 0K region: 0x727000000000
29 ( 2560): mapped: 256K allocs: 25 frees: 0 inuse: 25 num_freed_chunks 77 avail: 102 rss: 20K releases: 0 last released: 0K region: 0x727400000000
30 ( 3072): mapped: 256K allocs: 21 frees: 0 inuse: 21 num_freed_chunks 64 avail: 85 rss: 8K releases: 0 last released: 0K region: 0x727800000000
31 ( 3584): mapped: 256K allocs: 18 frees: 0 inuse: 18 num_freed_chunks 55 avail: 73 rss: 12K releases: 0 last released: 0K region: 0x727c00000000
32 ( 4096): mapped: 256K allocs: 16 frees: 0 inuse: 16 num_freed_chunks 48 avail: 64 rss: 12K releases: 0 last released: 0K region: 0x728000000000
33 ( 5120): mapped: 256K allocs: 12 frees: 0 inuse: 12 num_freed_chunks 39 avail: 51 rss: 16K releases: 0 last released: 0K region: 0x728400000000
34 ( 6144): mapped: 256K allocs: 10 frees: 0 inuse: 10 num_freed_chunks 32 avail: 42 rss: 12K releases: 0 last released: 0K region: 0x728800000000
35 ( 7168): mapped: 256K allocs: 9 frees: 0 inuse: 9 num_freed_chunks 27 avail: 36 rss: 52K releases: 0 last released: 0K region: 0x728c00000000
36 ( 8192): mapped: 256K allocs: 8 frees: 0 inuse: 8 num_freed_chunks 24 avail: 32 rss: 16K releases: 0 last released: 0K region: 0x729000000000
37 ( 10240): mapped: 256K allocs: 6 frees: 0 inuse: 6 num_freed_chunks 19 avail: 25 rss: 12K releases: 0 last released: 0K region: 0x729400000000
38 ( 12288): mapped: 256K allocs: 5 frees: 0 inuse: 5 num_freed_chunks 16 avail: 21 rss: 12K releases: 0 last released: 0K region: 0x729800000000
39 ( 14336): mapped: 256K allocs: 4 frees: 0 inuse: 4 num_freed_chunks 14 avail: 18 rss: 16K releases: 0 last released: 0K region: 0x729c00000000
42 ( 24576): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 8 avail: 10 rss: 16K releases: 0 last released: 0K region: 0x72a800000000
43 ( 28672): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 7 avail: 9 rss: 28K releases: 0 last released: 0K region: 0x72ac00000000
44 ( 32768): mapped: 256K allocs: 2 frees: 0 inuse: 2 num_freed_chunks 6 avail: 8 rss: 32K releases: 0 last released: 0K region: 0x72b000000000
45 ( 40960): mapped: 256K allocs: 1 frees: 0 inuse: 1 num_freed_chunks 5 avail: 6 rss: 4K releases: 0 last released: 0K region: 0x72b400000000
Stats: LargeMmapAllocator: allocated 1 times, remains 0 (0 K) max 1 M; by size logs: 20:1;
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ kill 1348
+ rg -Fa '' /var/log/clickhouse-server/clickhouse-server.log
2025.04.02 09:40:35.084346 [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Logical error: 'Replica info is not initialized'.
2025.04.02 09:40:35.160042 [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Stack trace (when copying this message, always include the lines below):
2025.04.02 09:40:35.162901 [ 2683 ] {} BaseDaemon: ########## Short fault info ############
2025.04.02 09:40:35.163404 [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2170) Received signal 6
2025.04.02 09:40:35.163744 [ 2683 ] {} BaseDaemon: Signal description: Aborted
2025.04.02 09:40:35.163823 [ 2683 ] {} BaseDaemon:
2025.04.02 09:40:35.163945 [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x000055701ef34a3c 0x000055701ef31d6e 0x000055701ef33c0d 0x00005570235a9bfe 0x0000557023089f8b 0x00005570235a984d 0x00005570230aea47 0x000055702309e54c 0x000055702309d8cc 0x000055702309d5aa 0x000055702309c36a 0x000055701706b7ee 0x00005570170739dc 0x000055700e684428 0x00007fd7c201fac3 0x00007fd7c20b1850
2025.04.02 09:40:35.164119 [ 2683 ] {} BaseDaemon: ########################################
2025.04.02 09:40:35.165045 [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2170) (query_id: 5495cad3-cad2-435c-8e6e-a56e3e787891) (query: INSERT INTO distributed_01099_b SELECT * FROM urlCluster('test_cluster_two_shards', 'http://localhost:8123/?query=select+{1,2,3}+format+TSV', 'TSV', 's String');) Received signal Aborted (6)
2025.04.02 09:40:35.165988 [ 2683 ] {} BaseDaemon:
2025.04.02 09:40:35.166563 [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x000055701ef34a3c 0x000055701ef31d6e 0x000055701ef33c0d 0x00005570235a9bfe 0x0000557023089f8b 0x00005570235a984d 0x00005570230aea47 0x000055702309e54c 0x000055702309d8cc 0x000055702309d5aa 0x000055702309c36a 0x000055701706b7ee 0x00005570170739dc 0x000055700e684428 0x00007fd7c201fac3 0x00007fd7c20b1850
2025.04.02 09:40:35.369611 [ 2683 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()
2025.04.02 09:40:35.369941 [ 2683 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d
2025.04.02 09:40:35.566929 [ 2683 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422
2025.04.02 09:40:41.719899 [ 2683 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6
2025.04.02 09:40:58.543986 [ 2683 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb
2025.04.02 09:40:58.550350 [ 2683 ] {} BaseDaemon: 4. ? @ 0x00007fd7c1fcd520
2025.04.02 09:40:58.552001 [ 2683 ] {} BaseDaemon: 5. ? @ 0x00007fd7c20219fd
2025.04.02 09:40:58.552734 [ 2683 ] {} BaseDaemon: 6. ? @ 0x00007fd7c1fcd476
2025.04.02 09:40:58.552991 [ 2683 ] {} BaseDaemon: 7. ? @ 0x00007fd7c1fb37f3
2025.04.02 09:41:18.915371 [ 2683 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267
2025.04.02 09:41:19.730227 [ 2683 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a
2025.04.02 09:41:19.946326 [ 2683 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)
2025.04.02 09:41:19.949528 [ 2683 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff
2025.04.02 09:41:40.626177 [ 2683 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade
2025.04.02 09:41:57.988457 [ 2683 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba
2025.04.02 09:42:02.121691 [ 2683 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c
2025.04.02 09:42:02.526694 [ 2683 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e
2025.04.02 09:42:02.984820 [ 2683 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d
2025.04.02 09:42:03.568477 [ 2683 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe
2025.04.02 09:42:03.880034 [ 2683 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b
2025.04.02 09:42:03.988427 [ 2683 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d
2025.04.02 09:42:04.260448 [ 2683 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)
2025.04.02 09:42:04.269401 [ 2683 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47
2025.04.02 09:42:05.376232 [ 2683 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c
2025.04.02 09:42:06.230286 [ 2683 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)
2025.04.02 09:42:06.231304 [ 2683 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc
2025.04.02 09:42:06.373556 [ 2683 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa
2025.04.02 09:42:06.671125 [ 2683 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)
2025.04.02 09:42:06.675053 [ 2683 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()
2025.04.02 09:42:06.675906 [ 2683 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025.04.02 09:42:06.676393 [ 2683 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE
2025.04.02 09:42:06.678745 [ 2683 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)
2025.04.02 09:42:06.680546 [ 2683 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()
2025.04.02 09:42:06.681280 [ 2683 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?
2025.04.02 09:42:06.681932 [ 2683 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?
2025.04.02 09:42:06.682611 [ 2683 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?
2025.04.02 09:42:06.685281 [ 2683 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a
2025.04.02 09:42:07.228552 [ 2683 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?
2025.04.02 09:42:07.231516 [ 2683 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?
2025.04.02 09:42:07.231787 [ 2683 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee
2025.04.02 09:42:07.591536 [ 2683 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?
2025.04.02 09:42:07.595342 [ 2683 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*, 2ul>(std::tuple>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>&, std::__tuple_indices<2ul>)
2025.04.02 09:42:07.596206 [ 2683 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl::ThreadFromThreadPool*>>(void*) @ 0x00000000104f79dc
2025.04.02 09:42:24.747786 [ 2683 ] {} BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428
2025.04.02 09:42:24.749564 [ 2683 ] {} BaseDaemon: 27. ? @ 0x00007fd7c201fac3
2025.04.02 09:42:24.749778 [ 2683 ] {} BaseDaemon: 28. ? @ 0x00007fd7c20b1850
2025.04.02 09:42:37.541566 [ 2683 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)
2025.04.02 09:43:54.468595 [ 2683 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.
2025.04.02 09:43:54.470745 [ 2683 ] {} BaseDaemon: Changed settings: min_compress_block_size = 1147731, max_compress_block_size = 2885936, max_block_size = 62186, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 98860, max_insert_threads = 2, max_threads = 1, max_read_buffer_size = 995864, connect_timeout_with_failover_ms = 2000, connect_timeout_with_failover_secure_ms = 3000, idle_connection_timeout = 36000, s3_max_get_rps = 1000000, s3_max_get_burst = 2000000, s3_max_put_rps = 1000000, s3_max_put_burst = 2000000, s3_check_objects_after_upload = true, use_uncompressed_cache = true, max_remote_read_network_bandwidth = 1000000000000, max_remote_write_network_bandwidth = 1000000000000, max_local_read_bandwidth = 1000000000000, max_local_write_bandwidth = 1000000000000, stream_like_engine_allow_direct_select = true, enable_multiple_prewhere_read_steps = false, replication_wait_for_inactive_replica_timeout = 30, min_count_to_compile_expression = 0, compile_aggregate_expressions = false, min_count_to_compile_aggregate_expression = 0, group_by_two_level_threshold = 881876, group_by_two_level_threshold_bytes = 13928924, enable_memory_bound_merging_of_aggregation_results = false, allow_nonconst_timezone_arguments = true, parallel_distributed_insert_select = 1, input_format_parallel_parsing = false, min_chunk_bytes_for_parallel_parsing = 17686868, merge_tree_coarse_index_granularity = 26, min_bytes_to_use_direct_io = 10737418240, min_bytes_to_use_mmap_io = 1, log_queries = true, insert_quorum_timeout = 60000, merge_tree_read_split_ranges_into_intersecting_and_non_intersecting_injection_probability = 0.5199999809265137, http_wait_end_of_query = true, http_response_buffer_size = 3622091, fsync_metadata = false, http_send_timeout = 60., http_receive_timeout = 60., use_index_for_in_with_subqueries_max_values = 1000000000, opentelemetry_start_trace_probability = 0.10000000149011612, max_rows_to_read = 20000000, max_bytes_to_read = 1000000000000, max_bytes_to_read_leaf = 1000000000000, max_rows_to_group_by = 10000000000, max_bytes_before_external_group_by = 10737418240, max_rows_to_sort = 10000000000, max_bytes_to_sort = 10000000000, prefer_external_sort_block_bytes = 0, max_bytes_before_external_sort = 10737418240, max_bytes_before_remerge_sort = 1309929691, max_result_rows = 1000000000, max_result_bytes = 1000000000, max_execution_time = 600., max_execution_time_leaf = 600., max_execution_speed = 100000000000, max_execution_speed_bytes = 10000000000000, timeout_before_checking_execution_speed = 300., max_estimated_execution_time = 600., max_columns_to_read = 20000, max_temporary_columns = 20000, max_temporary_non_const_columns = 20000, max_rows_in_set = 10000000000, max_bytes_in_set = 10000000000, max_rows_in_join = 10000000000, max_bytes_in_join = 10000000000, cross_join_min_rows_to_compress = 1, cross_join_min_bytes_to_compress = 0, max_rows_to_transfer = 1000000000, max_bytes_to_transfer = 1000000000, max_rows_in_distinct = 10000000000, max_bytes_in_distinct = 10000000000, max_memory_usage = 5000000000, max_memory_usage_for_user = 32000000000, max_untracked_memory = 1048576, memory_profiler_step = 1048576, max_network_bandwidth = 100000000000, max_network_bytes = 1000000000000, max_network_bandwidth_for_user = 100000000000, max_network_bandwidth_for_all_users = 100000000000, max_temporary_data_on_disk_size_for_user = 100000000000, max_temporary_data_on_disk_size_for_query = 100000000000, max_backup_bandwidth = 100000000000, log_comment = '01099_parallel_distributed_insert_select.sql', send_logs_level = 'error', prefer_localhost_replica = false, aggregation_in_order_max_block_bytes = 42302895, read_in_order_two_level_merge_threshold = 83, max_hyperscan_regexp_length = 1000000, max_hyperscan_regexp_total_length = 10000000, allow_introspection_functions = true, database_atomic_wait_for_drop_and_detach_synchronously = true, optimize_if_chain_to_multiif = true, optimize_if_transform_strings_to_enum = true, optimize_append_index = true, query_cache_max_size_in_bytes = 10000000, query_cache_max_entries = 100000, distributed_ddl_entry_format_version = 6, external_storage_max_read_rows = 10000000000, external_storage_max_read_bytes = 10000000000, remote_filesystem_read_method = 'read', merge_tree_min_bytes_per_task_for_remote_reading = 16777216, merge_tree_compact_parts_min_granules_to_multibuffer_read = 101, async_insert_busy_timeout_max_ms = 5000, enable_filesystem_cache = true, enable_filesystem_cache_on_write_operations = true, throw_on_error_from_cache_on_write_operations = true, filesystem_cache_segments_batch_size = 3, use_page_cache_for_disks_without_file_cache = true, page_cache_inject_eviction = true, load_marks_asynchronously = true, allow_prefetched_read_pool_for_remote_filesystem = true, allow_prefetched_read_pool_for_local_filesystem = false, filesystem_prefetch_step_marks = 50, filesystem_prefetch_max_memory_usage = 67108864, filesystem_prefetches_limit = 10, max_streams_for_merge_tree_reading = 1000, optimize_sorting_by_input_stream_properties = false, insert_keeper_max_retries = 100, insert_keeper_retry_initial_backoff_ms = 1, insert_keeper_retry_max_backoff_ms = 10, insert_keeper_fault_injection_probability = 0.009999999776482582, optimize_distinct_in_order = false, allow_experimental_parallel_reading_from_replicas = 0, session_timezone = 'Asia/Anadyr', output_format_native_write_json_as_string = true
2025.04.02 09:43:58.819850 [ 430 ] {} Application: Child process was terminated by signal 6.
+ rg -A50 -Fa ============ /var/log/clickhouse-server/stderr.log
+ :
+ data_path_config=--path=/var/lib/clickhouse/
+ [[ -n '' ]]
+ zstd --threads=0
+ '[' 1 -ne 0 ']'
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.zookeeper_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.trace_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ zstd --threads=0
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.transactions_info_log format TSVWithNamesAndTypes'
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.metric_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.blob_storage_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.error_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for table in query_log zookeeper_log trace_log transactions_info_log metric_log blob_storage_log error_log query_metric_log
+ clickhouse-local --max_block_size 8192 --max_memory_usage 10G --max_threads 1 --max_result_rows 0 --max_result_bytes 0 --max_bytes_to_read 0 --path=/var/lib/clickhouse/ --only-system-tables --stacktrace -q 'select * from system.query_metric_log format TSVWithNamesAndTypes'
+ zstd --threads=0
+ [[ 0 -eq 1 ]]
+ [[ 0 -eq 1 ]]
+ for trace_type in CPU Memory Real
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''CPU'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ zstd --threads=0
+ for trace_type in CPU Memory Real
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''Memory'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ zstd --threads=0
+ for trace_type in CPU Memory Real
+ zstd --threads=0
+ clickhouse-local --path=/var/lib/clickhouse/ --only-system-tables -q '
select
arrayStringConcat((arrayMap(x -> concat(splitByChar('\''/'\'', addressToLine(x))[-1], '\''#'\'', demangle(addressToSymbol(x)) ), trace)), '\'';'\'') AS stack,
count(*) AS samples
from system.trace_log
where trace_type = '\''Real'\''
group by trace
order by samples desc
settings allow_introspection_functions = 1
format TabSeparated'
+ check_logs_for_critical_errors
+ sed -n '/WARNING:.*anitizer/,/^$/p' /var/log/clickhouse-server/stderr.log
+ rg -Fav -e 'ASan doesn'\''t fully support makecontext/swapcontext functions' -e DB::Exception /test_output/tmp
+ echo -e 'No sanitizer asserts\tOK\t\N\t'
+ rm -f /test_output/tmp
+ rg -Fa ' Application: Child process was terminated by signal 9' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'No OOM messages in clickhouse-server.log\tOK\t\N\t'
+ rg -Fa 'Code: 49. DB::Exception: ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'No logical errors\tOK\t\N\t'
+ '[' -s /test_output/logical_errors.txt ']'
+ rm /test_output/logical_errors.txt
+ rg --text 'Code: 499.*The specified key does not exist' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ grep -v -e a.myext -e DistributedCacheTCPHandler -e ReadBufferFromDistributedCache -e ReadBufferFromS3 -e ReadBufferFromAzureBlobStorage -e AsynchronousBoundedReadBuffer -e 'caller id: None:DistribCache'
+ echo -e 'No lost s3 keys\tOK\t\N\t'
+ rg -Fa 'it is lost forever' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ grep SharedMergeTreePartCheckThread
+ echo -e 'No SharedMergeTree lost forever in clickhouse-server.log\tOK\t\N\t'
+ '[' -s /test_output/no_such_key_errors.txt ']'
+ rm /test_output/no_such_key_errors.txt
+ rg -Fa '#######################################' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
+ echo -e 'Killed by signal (in clickhouse-server.log)\tFAIL\t\N\t'
+ rg -Fa ' ' /var/log/clickhouse-server/clickhouse-server.err.log /var/log/clickhouse-server/clickhouse-server.log
++ trim_server_logs fatal_messages.txt
++ head -n 100 /test_output/fatal_messages.txt
++ grep -Eo ' \[ [0-9]+ \] \{.*'
++ escaped
++ clickhouse local -S 's String' --input-format=LineAsString -q 'select substr(s, 1, 300)
from table format CustomSeparated settings format_custom_row_after_delimiter='\''\\\\n'\'''
+ echo -e 'Fatal message in clickhouse-server.log (see fatal_messages.txt)\tFAIL\t\N\t [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Stack trace (when copying this message, always include the lines below):\\n [ 2683 ] {} BaseDaemon: ########## Short fault info ############\\n [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2170) Received signal 6\\n [ 2683 ] {} BaseDaemon: Signal description: Aborted\\n [ 2683 ] {} BaseDaemon: \\n [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x00005\\n [ 2683 ] {} BaseDaemon: ########################################\\n [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2170) (query_id: 5495cad3-cad2-435c-8e6e-a56e3e787891) (query: INSERT INTO distributed_01099_\\n [ 2683 ] {} BaseDaemon: \\n [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x00005\\n [ 2683 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 2683 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 2683 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 2683 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 2683 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 2683 ] {} BaseDaemon: 4. ? @ 0x00007fd7c1fcd520\\n [ 2683 ] {} BaseDaemon: 5. ? @ 0x00007fd7c20219fd\\n [ 2683 ] {} BaseDaemon: 6. ? @ 0x00007fd7c1fcd476\\n [ 2683 ] {} BaseDaemon: 7. ? @ 0x00007fd7c1fb37f3\\n [ 2683 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 2683 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 2683 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 2683 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 2683 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 2683 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 2683 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 2683 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 2683 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 2683 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 2683 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 2683 ] {} BaseDaemon: 18. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:134: DB::RemoteSource::work() @ 0x000000001ca2d84d\\n [ 2683 ] {} BaseDaemon: 19.0. inlined from ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:49: DB::executeJob(DB::ExecutingGraph::Node*, DB::ReadProgressCallback*)\\n [ 2683 ] {} BaseDaemon: 19. ./build_docker/./src/Processors/Executors/ExecutionThreadContext.cpp:98: DB::ExecutionThreadContext::executeTask() @ 0x000000001c532a47\\n [ 2683 ] {} BaseDaemon: 20. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:290: DB::PipelineExecutor::executeStepImpl(unsigned long, std::atomic*) @ 0x000000001c52254c\\n [ 2683 ] {} BaseDaemon: 21.0. inlined from ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:256: DB::PipelineExecutor::executeSingleThread(unsigned long)\\n [ 2683 ] {} BaseDaemon: 21. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:442: DB::PipelineExecutor::executeImpl(unsigned long, bool) @ 0x000000001c5218cc\\n [ 2683 ] {} BaseDaemon: 22. ./build_docker/./src/Processors/Executors/PipelineExecutor.cpp:127: DB::PipelineExecutor::execute(unsigned long, bool) @ 0x000000001c5215aa\\n [ 2683 ] {} BaseDaemon: 23.0. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:49: DB::threadFunction(DB::CompletedPipelineExecutor::Data&, std::shared_ptr, unsigned long, bool)\\n [ 2683 ] {} BaseDaemon: 23.1. inlined from ./build_docker/./src/Processors/Executors/CompletedPipelineExecutor.cpp:89: operator()\\n [ 2683 ] {} BaseDaemon: 23.2. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 2683 ] {} BaseDaemon: 23.3. inlined from ./contrib/llvm-project/libcxx/include/tuple:1789: _ZNSt3__118__apply_tuple_implB6v15007IRZN2DB25CompletedPipelineExecutor7executeEvE3$_0RNS_5tupleIJEEETpTnmJEEEDcOT_OT0_NS_15__tuple_indicesIJXspT1_EEEE\\n [ 2683 ] {} BaseDaemon: 23.4. inlined from ./contrib/llvm-project/libcxx/include/tuple:1798: decltype(auto) std::apply[abi:v15007]&>(DB::CompletedPipelineExecutor::execute()::$_0&, std::tuple<>&)\\n [ 2683 ] {} BaseDaemon: 23.5. inlined from ./src/Common/ThreadPool.h:311: operator()\\n [ 2683 ] {} BaseDaemon: 23.6. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:394: ?\\n [ 2683 ] {} BaseDaemon: 23.7. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:479: ?\\n [ 2683 ] {} BaseDaemon: 23.8. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:235: ?\\n [ 2683 ] {} BaseDaemon: 23. ./contrib/llvm-project/libcxx/include/__functional/function.h:716: ? @ 0x000000001c52036a\\n [ 2683 ] {} BaseDaemon: 24.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:848: ?\\n [ 2683 ] {} BaseDaemon: 24.1. inlined from ./contrib/llvm-project/libcxx/include/__functional/function.h:1197: ?\\n [ 2683 ] {} BaseDaemon: 24. ./build_docker/./src/Common/ThreadPool.cpp:785: ThreadPoolImpl::ThreadFromThreadPool::worker() @ 0x00000000104ef7ee\\n [ 2683 ] {} BaseDaemon: 25.0. inlined from ./contrib/llvm-project/libcxx/include/__functional/invoke.h:359: ?\\n [ 2683 ] {} BaseDaemon: 25.1. inlined from ./contrib/llvm-project/libcxx/include/thread:284: void std::__thread_execute[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl\\n [ 2683 ] {} BaseDaemon: 25. ./contrib/llvm-project/libcxx/include/thread:295: void* std::__thread_proxy[abi:v15007]>, void (ThreadPoolImpl::ThreadFromThreadPool::*)(), ThreadPoolImpl BaseDaemon: 26. __tsan_thread_start_func @ 0x0000000007b08428\\n [ 2683 ] {} BaseDaemon: 27. ? @ 0x00007fd7c201fac3\\n [ 2683 ] {} BaseDaemon: 28. ? @ 0x00007fd7c20b1850\\n [ 2683 ] {} BaseDaemon: Integrity check of the executable successfully passed (checksum: CAD5C347B3B532C05434FCAB6F84FB12)\\n [ 2683 ] {} BaseDaemon: This ClickHouse version is not official and should be upgraded to the official build.\\n [ 2683 ] {} BaseDaemon: Changed settings: min_compress_block_size = 1147731, max_compress_block_size = 2885936, max_block_size = 62186, min_external_table_block_size_bytes = 0, max_joined_block_size_rows = 98860, max_insert_threads = 2, max_threads = 1, max_read_buffer_size = 995864, connec\\n [ 430 ] {} Application: Child process was terminated by signal 6.\\n [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Logical error: \'\''Replica info is not initialized\'\''.\\n [ 2170 ] {5495cad3-cad2-435c-8e6e-a56e3e787891} : Stack trace (when copying this message, always include the lines below):\\n [ 2683 ] {} BaseDaemon: ########## Short fault info ############\\n [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d, architecture: x86_64) (from thread 2170) Received signal 6\\n [ 2683 ] {} BaseDaemon: Signal description: Aborted\\n [ 2683 ] {} BaseDaemon: \\n [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x00005\\n [ 2683 ] {} BaseDaemon: ########################################\\n [ 2683 ] {} BaseDaemon: (version 24.12.2.20221.altinityantalya (altinity build), build id: 1D546DEA3F7DAB479A2081B78470431AC256AC96, git hash: 82252d159dc02cab0f366aaa5691adc1545dd11d) (from thread 2170) (query_id: 5495cad3-cad2-435c-8e6e-a56e3e787891) (query: INSERT INTO distributed_01099_\\n [ 2683 ] {} BaseDaemon: \\n [ 2683 ] {} BaseDaemon: Stack trace: 0x0000557016fa011d 0x000055701739d422 0x000055700e68c7a6 0x000055700e68ccfb 0x00007fd7c1fcd520 0x00007fd7c20219fd 0x00007fd7c1fcd476 0x00007fd7c1fb37f3 0x000055700e68b267 0x0000557016f58a0a 0x0000557016f59dff 0x000055700e711ade 0x000055700e7219ba 0x00005\\n [ 2683 ] {} BaseDaemon: 0.0. inlined from ./build_docker/./src/Common/StackTrace.cpp:381: StackTrace::tryCapture()\\n [ 2683 ] {} BaseDaemon: 0. ./build_docker/./src/Common/StackTrace.cpp:350: StackTrace::StackTrace(ucontext_t const&) @ 0x000000001042411d\\n [ 2683 ] {} BaseDaemon: 1. ./build_docker/./src/Common/SignalHandlers.cpp:102: signalHandler(int, siginfo_t*, void*) @ 0x0000000010821422\\n [ 2683 ] {} BaseDaemon: 2. __tsan::CallUserSignalHandler(__tsan::ThreadState*, bool, bool, int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b107a6\\n [ 2683 ] {} BaseDaemon: 3. sighandler(int, __sanitizer::__sanitizer_siginfo*, void*) @ 0x0000000007b10cfb\\n [ 2683 ] {} BaseDaemon: 4. ? @ 0x00007fd7c1fcd520\\n [ 2683 ] {} BaseDaemon: 5. ? @ 0x00007fd7c20219fd\\n [ 2683 ] {} BaseDaemon: 6. ? @ 0x00007fd7c1fcd476\\n [ 2683 ] {} BaseDaemon: 7. ? @ 0x00007fd7c1fb37f3\\n [ 2683 ] {} BaseDaemon: 8. ___interceptor_abort @ 0x0000000007b0f267\\n [ 2683 ] {} BaseDaemon: 9. ./build_docker/./src/Common/Exception.cpp:48: DB::abortOnFailedAssertion(String const&, void* const*, unsigned long, unsigned long) @ 0x00000000103dca0a\\n [ 2683 ] {} BaseDaemon: 10.0. inlined from ./build_docker/./src/Common/Exception.cpp:70: DB::handle_error_code(String const&, int, bool, std::vector> const&)\\n [ 2683 ] {} BaseDaemon: 10. ./build_docker/./src/Common/Exception.cpp:111: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x00000000103dddff\\n [ 2683 ] {} BaseDaemon: 11. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000007b95ade\\n [ 2683 ] {} BaseDaemon: 12. DB::Exception::Exception<>(int, FormatStringHelperImpl<>) @ 0x0000000007ba59ba\\n [ 2683 ] {} BaseDaemon: 13. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:727: DB::RemoteQueryExecutor::processReadTaskRequest() @ 0x00000000183b8a3c\\n [ 2683 ] {} BaseDaemon: 14. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:623: DB::RemoteQueryExecutor::processPacket(DB::Packet) @ 0x00000000183b5d6e\\n [ 2683 ] {} BaseDaemon: 15. ./build_docker/./src/QueryPipeline/RemoteQueryExecutor.cpp:562: DB::RemoteQueryExecutor::readAsync() @ 0x00000000183b7c0d\\n [ 2683 ] {} BaseDaemon: 16. ./build_docker/./src/Processors/Sources/RemoteSource.cpp:182: DB::RemoteSource::tryGenerate() @ 0x000000001ca2dbfe\\n [ 2683 ] {} BaseDaemon: 17. ./build_docker/./src/Processors/ISource.cpp:108: DB::ISource::work() @ 0x000000001c50df8b\\n [ 2683 ] {}